Category Archives: Uncategorized

My recent COVID-19 commentary . . .

Information is the key to defeating COVID-19

The chief constraints on the world are physical and temporal. Abundant information, however, helps us evade and sometimes conquer these facts of life. Without the internet, life in the middle of the COVID-19 pandemic would be far nastier and more brutish, even more solitary, and shorter, too. 

Medical staff arrive with a patient affected by COVID-19 at a hospital in Liege, Belgium, March 23, 2020 – via REUTERS

In our current predicament, the fundamental shortcomings of space and time are amplified. SARS-CoV-2 can’t replicate without our cells and our proximity to each other. So we separate for a period longer than the viral lifetime. Yet our bodies need calories, and time does not stop for debts and rents to be paid. 

So we must buy time with information

Extreme physical dis-connectivity requires extreme virtual connectivity. The speed of the virtual scale-up has been impressive. Teachers have adroitly shifted to meet with 30 million school children online. Colleges the same. Zoom video conferencing has spread faster than the virus, and Microsoft Teams added 12 million users in just one week, growing by more than a third to 44 million. Communities are leveraging social networks to support neighbors in a crunch. The government is (finally!) allowing doctors and patients to use FaceTime and Skype for remote video check-ups. The American internet infrastructure has handled increased demand admirably, with internet service providers and mobile carriers adding capacity and suspending data usage limits. 

Not every activity can be replicated online, however. And so we call on the wealth generated by our information-rich economy. Cash infusions, liquidity, loans, and forbearance can smooth away the sudden halt of face-to-face work — at least for a time. Massive investments in hospitals and medical supplies will save many of today’s ill, and tomorrow’s too. Wealth in a very real sense is resilience

But wealth must be constantly regenerated. We can only rely on past productivity for so long. And so we must get back to work — as quickly as possible, in as many places as possible. 

From macro to micro

We must replace our crucial macro response with an equally crucial micro recovery. Our initial macro efforts were broad, blunt, and indiscriminate. And understandably so. We didn’t have enough information to finely tune closings and distancing for particular people and places, nor to support earlier, more comprehensive travel bans, which would have been even more controversial. God willing, our macro efforts will spare large parts of the nation from intense breakouts of the type ravaging Milan, New York, and Seattle. 

We need to think about the next phase, however. To fully support our heroic medical community in hard-hit places and vulnerable Americans everywhere, we must quickly reignite as much of the economy as possible. And we can only do so with better information. 

Our micro efforts to target people, places, and things, and to collect exabytes of data, will be central. Which people and places are safe to return to work and school? For that, we need widespread testing. Which travel routes are safe? Which hospitals need (or can spare) extra capacity and supplies? Which early off-label and experimental therapies are showing the most promise? What are the immunity characteristics of COVID-19? Who contracted it without knowing it? And how will these answers inform our immunity strategy for any second wave this fall?

Better information can support a strategy of smart engagement. Without it, blunt macro policies will prevent the agility necessary in the days ahead. These efforts will require a heroic scale-up of information gathering tools and ideas. Most of them will not come from government (which will need to play a supporting and coordinating role), but from private firms and organizations. 

The app-ification of medicine

We can launch a new era in radically decentralized personal medicine — for better individual health, an explosion in physician productivity, a research renaissance into new therapies, and far better public health surveillance. 

In the future, massive data will detect outbreaks and smash them early, and most of the world economy can go on while risk zones are isolated and treated. Surveillance does not mean government watching your sneezes or temperature. It means mostly anonymous data collection, perhaps by third parties who can detect outbreaks and issue alerts. The goal is not zero cases. Such a goal would turn the government into an authoritarian police state. To protect both public health and private liberties, this is going to have to be a true public-private partnership.

All these things will require the FDA, CDC, and other federal agencies to adopt a new ethos of innovation, or to get out of the way. It will not come naturally. But these events must shake us out of our dangerous complacencies: the CDC’s faulty and thus delayed test; the FDA’s initial resistance to off-label and experimental therapies; the FDA’s reported resistance to Apple incorporating more health sensors and apps into the iPhone and Apple Watch.

A new bio boom The ultimate information tool is our understanding of the genome and all the “omics.” Although young, computational biology and related fields are now advancing at an astounding pace. They helped decode SARS-CoV-2’s genetic sequence in record time and, God willing, will help deliver a vaccine in record time as well. These codes of life can, along with our information technologies, lead to healthier bodies and more invaluable time with each other.

This article originally appeared at AEIdeas.

A Plan for Anti-Viral Victory

Sensing despair in the moment, I outlined a path to recovery. Originally on Twitter . . . .

A PLAN for ANTI-VIRAL VICTORY — Leveraging information to overcome our physical and temporal challenges

Health & economy are intertwined. After much-needed adoption of major behavioral modifications, we now should prepare for pivot to Rapid Recovery. — Bret Swanson (@JBSay) March 20, 2020

Prolonged economic paralysis will undermine crucial health efforts. We can creatively leverage our extraordinary broadband information infrastructure for much of the economy. But we should restart the physical economy in as many places as possible, as soon as possible.

Encouraging news that pharma firms are providing big supplies – tens of millions of pills – of off-label drugs (eg HCQ) that *appear* to be effective vs. Covid-19. More speed on more fronts will help immediately – AND set new innovation-tilting precedents.

We can also launch a new era in personal medicine, with massive data so that future outbreaks can be detected and smashed early, and so most of the world economy can go on while Risk Zones are isolated and treated.

Surveillance does not mean government watching your sneezes or temperature. It means anon data collection, perhaps by third parties who can detect outbreaks and issue alerts. The goal is not ZERO cases. Such a goal would turn the government into an authoritarian police state.

Our problems are chiefly physical and temporal. We can use information to mitigate some physical and temporal dislocations, and for those we can’t, we will call on the wealth of our info-rich economy to smooth away the viral lifetime.

A short-term silver lining can be a new focus on teamwork, shared humanity, and practicality over pettiness.

A medium-term silver lining could be speedier adoption by the physical economy of information tools that can make them more productive and creative.

A long-term silver lining can be a new commitment to radical bio innovation and a transformed (better/less expensive) health sector, an acceleration of the App-ification of Medicine.

McKitrick’s wisdom on the state of the climate debate

Must reading from the excellent environmental economist Ross McKitrick:

“The old compromise is dead. Stop using C jargon in your speeches. Start learning the deep details of the science and economics instead of letting the C crowd dictate what you’re allowed to think or say. Figure out a new way of talking about the climate issue based on what you actually believe. Learn to make the case for Canada’s economy to survive and grow.

“You, and by extension everyone who depends on your leadership, face an existential threat. It was 20 years in the making, so dig in for a 20-year battle to turn it around. Stop demonizing potential allies in the A camp; you need all the help you can get.

“Climate and energy policy has fallen into the hands of a worldwide movement that openly declares its extremism. The would-be moderates on this issue have pretended for 20 years they could keep the status quo without having to fight for it. Those days are over.”

Seizing the 5G mid-band spectrum opportunity

Tomorrow, the House Energy and Commerce Committee will hear testimony on one of the spectrum bands most important to the future of 5G wireless. Known as the C-Band, the frequencies are located between 3.7 and 4.2 gigahertz (GHz), which is a kind of sweet spot in the fundamental tradeoff between distance and data rates for all wireless technologies. The spectrum is perfect for the new 5G architecture, which will rely on hundreds of thousands or even millions of small cells, already popping up across the country and around the world.

Right now, however, this band is used by three satellite firms (Intelsat, SES, and Telesat) to broadcast video and radio content to cable TV and radio ground stations for further delivery via terrestrial networks to consumers. Technology is changing this market, however. Increasingly, video and radio content is delivered to distribution points via fiber optic lines, not satellite, and so the satellite firms are left with extra spectrum, and they’d like to sell it. Which works out perfectly. Because the mobile firms would love to buy it for 5G.

The spectrum is not only perfectly situated for small cell deployments, but there’s also a lot of it. Out of the 500 MHz total held by the satellite firms today, potentially 300 MHz could be auctioned to the mobile firms. To get an idea of how much 300 MHz is, consider that today in the U.S. the mobile firms operate their networks with a deployed spectrum total of only around 580 MHz. So the C band could more than double today’s mobile airwaves.

That’s great news. But moving this spectrum from one set of users and technologies to another can be tricky. A group called the C Band Alliance has proposed a novel mechanism where the satellite firms would, under the supervision of the Federal Communications Commission (FCC), auction a big portion of the 500 MHz directly to the bidding mobile firms. They think the process of auctioning, repacking, and deploying the spectrum, along with necessary ground station and fiber optic upgrades, could be accomplished in two to three years.

Other firms, however, favor a series of incentive auctions run by the FCC. This second group claims the auctions could be done quickly. But that would belie all experience, where auctions this complex would likely take at least several more years, or more than double the time, of the C Band Alliance proposal.

It’s impossible to predict exactly how each proposal would work, and how long it would take. But given the importance of 5G to U.S. economy, at this point I favor speed, which means moving ahead with the C Band Alliance proposal.

For more on 5G, mid-band spectrum, and the economic implications, see some of our other recent items:

Filling the mid-band spectrum gap to sustain 5G momentum

Spectrum big bang points to promise of 5G wireless

Mobile wireless: A still under-appreciated economic miracle

If your content doesn’t appear on Google or Twitter, do you exist?

In the world of attention-getting, getting banished or downgraded on the world’s key attention platforms is frustrating. Or worse. The latest example is a claim of invisibility by Democratic presidential candidate Tulsi Gabbard. She claims that after the June presidential debate, just as she was gaining steam from a solid performance, Google deactivated the ads she had purchased on its platform, blocking people from finding her content when they searched for her. 

In recent days, most of the claims of blocking, throttling, shadow-banning, and demonetization on digital platforms has come from conservatives. Gabbard is a progressive, but some wonder whether her status as an outsider – as many of the conservatives are – is even more central to the seeming trend of suppression. Gabbard is suing Google for $50 million. She says Google’s excuses for why it deactivate her account for several crucial hours, before reactivating it, don’t add up. Much of her legal brief is couched in the Constitution. 

To this non-lawyer, Gabbard’s complaint seems rather weak. At least legally. The First Amendment protects Americans from government encroachments on speech, religion, and assembly. The Constitution doesn’t guarantee citizens the positive right to be heard on private platforms. What’s more, the big tech firms enjoy their own First Amendment rights. 

And yet, just because Gabbard’s complaint is feeble with regard to the First Amendment doesn’t mean it lacks substance as an example of a very real problem on the Internet. The blocking and throttling of Internet data based on viewpoints may not violate the Constitution, but it does violate our sense of fairness and our idea of what the Internet should be – a generally free and open platform for communication and content. 


Apollo, mankind, and Moore’s law

When Neil Armstrong, Buzz Aldrin, and Michael Collins went to the moon 50 years ago this week, they had a large portion of the world’s computing power with them on Columbia and Eagle and behind them in Houston. One NASA engineer estimated that, between 1962 and 1967, the Apollo program had purchased 60 percent of all integrated circuits built in the US.

An image of the Saturn V rocket, which launched the Apollo 11 astronauts into space, is projected onto the side of the Washington Monument to mark the 50th anniversary of the first lunar mission in Washington, DC July 16, 2019 – via REUTERS

Today, however, that overwhelming proportion seems paltry in its aggregate power. The two Apollo Guidance Computers (AGC) onboard the spacecraft, for example, each contained 32 kilobits of random-access memory and 72 kilobytes of read-only memory. The AGCs had a primary clock running at 2.048 megahertz, and their 2,048 integrated circuits contained only several tens of thousands of transistors. They also weighed 70 pounds.

By comparison, today’s iPhone XS sports 32 gigabits of dynamic random-access memory, 256 gigabytes of storage, and a processor with 6.9 billion transistors running at 2.49 gigahertz. That’s a million times more memory, several million times more storage, and hundreds of millions times more processing power than the AGCs. All in a package one-hundredth the weight.

Even in the late 1960s, however, NASA had begun enjoying the early fruits of Moore’s law. As The Wall Street Journal noted in one of its many impressive articles on Apollo 11’s anniversary, “the first computer chips tested by MIT” — which built the AGCs — “cost $1,000 each. By the time astronauts landed on the moon, the price had dropped to $15 apiece. . . . It set a pattern of innovation, quality control and price-cutting that persists in the computer business to this day.”

In another article about the software team, which had to do so much with such limited hardware, The Wall Street Journal described the tense moments just before landing on the moon, when the mission was nearly aborted:

Neil Armstrong hovered a few miles above the surface of the moon on July 20, 1969, searching for a safe place to make history.

Only minutes of fuel remained to land the first men on another world. A power meter failed in Apollo 11’s cramped lunar lander. Communications faded in and out. Then, warnings began flashing: Program alarm. Program alarm.

Five times the onboard computer signaled an emergency like none Armstrong and crewmate Buzz Aldrin had practiced.

In that moment, the lives of two astronauts, the efforts of more than 300,000 technicians, the labor of eight years at a cost of $25 billion, and the pride of a nation depended on a few lines of pioneering computer code.

Read the entire article and the related content. And of course watch last year’s feature film “First Man” and this year’s documentary “Apollo 11.” Looking back, Apollo 11’s skimpy digital capacity only amplifies the engineers’ creativity and genius and the astronauts’ bravery. With today’s technical capabilities and a little of their vision and determination, who knows what giant leaps are possible?

This item was originally published at AEIdeas.

Deplatforming and disinformation will degrade our democracy

We spent much of the last two decades talking about ways to expand access to information — boosting broadband speeds, extending mobile coverage, building Wikipedia and Google and Github. But now that the exafloods have washed over us, with more waves on the way, many of our new challenges are the result of information overload.

In a world of digital overabundance, how do we protect our privacy and our children’s innocence? How do we highlight the important stuff and block the nonsense? How do we filter, sort, safeguard, and verify?

Information is the currency of a culture and the basis of learning and growth. The information explosion has enriched us innumerably. But if we don’t successfully grapple with some of the downsides, we will forfeit this amazing gift.

Two of today’s biggest threats are disinformation (the spreading of false or misleading content) and deplatforming (blocking access to or manipulating some information hub). Neither is entirely new. But in our networked world, both effects are supercharged, and they strike at the heart of our society’s ability to process information effectively.

The founders thought our democratic experiment required an educated and informed citizenry. The growth of our experiment, likewise, requires the ability to generate new knowledge, which, in turn, requires disagreement, debate, and creativity. Without a grasp of reality and good faith efforts to generate new knowledge, however, the system can founder.

Over the last two years, we heard much about foreign disinformation campaigns targeting the 2016 election. But we’ve now learned that many of these alarmist charges were themselves elaborate disinformation campaigns. Fraudulent documents were pumped into our law enforcement agencies and sprinkled across the government and media. The social media tools used in modest, mostly ineffective ways by Russian trolls were then repurposed in the 2017 and 2018 elections by American political groups posing as anti-disinformation scientists. The self-described investigators of disinformation have in fact become the purveyors of disinformation.

A rational response to spam, vice, disinformation, or mere poor quality is to filter, sort, and prioritize. Thus institutions of all kinds make legitimate decisions to carry or disallow content or activity on their platforms.

Apple, for example, keeps drugs, gambling, and sex, among other vices, off of its App Store. That’s a perfectly good strategy for Apple, its customers, and society. Platforms have the right to develop their own product and culture. Some form of gatekeeping or prioritization at some nodes of our shifting networks will always be necessary.

Our sense of fairness, however, is offended when a supposedly open platform makes arbitrary or outright discriminatory decisions. If, for example, a platform doesn’t want to host political or scientific discussions, fine. But when a platform pretends to host broad ranges of content, including political, social, and scientific debate, then we expect some measure of neutrality.

In the last few years, however, these hubs have increasingly been captured by political activists, their own internal ideologies, or go-with-the-flow fads. Social networks, content repositories, and now even payment networks are deplatforming content and people deemed socially unacceptable. Many of those kicked off the platforms are thoroughly despicable characters, for sure. But mainstream activists, academics, thinkers, and even rival platforms are increasingly getting blocked, shadow banned, or otherwise suppressed by, for example, Twitter, YouTube, Facebook, Patreon, and PayPal. In fact, disinformation campaigns are a common way that rival activists get the hubs to deplatform enemies.

These tactics aren’t unique to the internet, of course. Disinformation is as old as time, or at least human warfare. And deplatforming is an unsavory trend on university campuses and academic journals, which are information hubs of a sort. Networks, however, amplify the power of these tactics. And so the new strategy of ideological badgering can also be found, and is especially potent, at large network nodes. Thus BlackRock, the $6-trillion family of index funds, which owns large percentages of all publicly traded companies, has become one of the activists’ juiciest targets. Instead of heckling every public firm or pension fund to do their political bidding, the activists successfully lobbied BlackRock to establish a “Stewardship Committee” to enforce their views, thus gaining some measure of control over all public firms without ownership of the firms.

The most astonishing current case is perhaps the most dangerous. It involves the apparent abuse of the government’s most powerful and sensitive surveillance tools and databases – the ultimate information platform — by political actors for political ends.

One structural solution to the politicization of centralized incumbents is to build rival institutions and decentralized platforms. arXiv is an alternative to traditional academic publishing, for example, and alternative news outlets continue to proliferate. Crypto- or blockchain-based peer-to-peer systems may be another way of disempowering the politicized platforms.

The current digital platforms might still regain some measure of public trust by recommitting to political and scientific neutrality (and to privacy, etc.). If they don’t, however, rival platforms will only grow faster. Washington, meanwhile, will be even more tempted to step in to regulate who can speak, what they can say, when, where, and how. In a misguided effort to bolster outcast speakers, free speech, a foundation of our system and our nation, will in fact likely suffer.

Some information platforms, such as law enforcement and intelligence, however, will inevitably remain unrivaled and centralized. And here we need a recommitment to professionalism, nonpartisanship, adult judgment, and farsighted citizenship.

Instead, our leadership class over the last many years has been a profound embarrassment. Perhaps poisoned by information overdose, government officials, public intellectuals, and journalists in their 50s, 60s, and 70s have behaved like the worst combination of toddlers and teenagers, gullible and paranoid, narrow-minded and spiteful. Supposedly educated and civilized men and women go on years-long rants on cable TV, while straight news has descended to its least accurate point ever. The commoditization of “the facts” has paradoxically expanded the field for factless nonsense.

What’s worse, closing off the spaces for rational inquiry will only deepen the social vertigo and prevent the course corrections needed to regain our individual and social balance. Despite all the real technological solutions to the challenge of information overload, human leadership and loftier social expectations may prove most important. To regain our balance, we desperately need our powers of science and civic discussion. Hold the current bad actors accountable, yes. But then we need to deescalate. Be skeptical — and invite skepticism of ourselves. Curate platforms for quality, but do not spitefully or ideologically discriminate. Be tough but generous and open.

The country needs robust information tools to defend itself and promote freedom across the globe. By information tools, I mean not just our military and intelligence capabilities. I refer also to free speech, science, and our open society. We cannot survive if these awesome powers are politicized and polluted.

This post originally appeared at AEIdeas –

5G wireless, fact and fiction

New wireless technologies, including 5G, are poised to expand the reach and robustness of mobile connectivity and boost broadband choices for tens of millions of consumers across the country. We’ve been talking about the potential of 5G the last few years, and now we are starting to see the reality. In a number of cities, thousands of small cells are going up on lampposts, utility poles, and building tops. I’ve discussed our own progress here in Indiana.

The project will take many years, but it’s happening. And the Federal Communications Commission just gave this massive infrastructure effort a lift by streamlining the rules for deploying these small cells. Because of the number of small cells to be deployed – many hundreds of thousands across the country – it would be counterproductive to treat each one of them as a new large structure, such as a building or hundred-foot cell tower. The new rules thus encourage fast deployment by smoothing the permitting process and making sure cities and states don’t charge excessive fees. The point is faster deployment of powerful new wireless networks, which will not only supercharge your smartphone but also provide a competitive alternative to traditional wired broadband.

Given this background, I found last week’s editorial by the mayor of San Jose, California, quite odd. Writing in the New York Times, Mayor Sam Liccardo argued that the new FCC rules to encourage faster deployment are an industry effort to “usurp control over these coveted public assets and utilize publicly owned streetlight poles for their own profit, not the public benefit.”

But the new streamlining rules do no such thing. Public rights of way will still be public. Cities and states will still have the same access as private firms, just as they had before. And who will benefit by the private investment of some $275 billion dollars in new wireless networks? That’s right – the public.

If cities and states wish to erect new Wi-Fi networks, as Mayor Liccardo did in San Jose, they can still do so.

I think the real complaint from some mayors is that the new FCC rules will limit their ability to extort wildly excessive fees and other payments from firms who want to bring these new wireless technologies to consumers. Too often, cities are blocking access to these rights of way, unless firms pay up. These government games are the very obstacles to deployment that the FCC rule is meant to fix.

Fewer obstacles, faster deployment. And accelerated deployment of the new 5G networks will mean broader coverage, faster speeds, and more broadband competition, which, crucially, will put downward pressure on connectivity prices, boosting broadband availability and affordability.

Mayor Liccardo emphasizes the challenges of low-income neighborhoods. But there are much better ways to help targeted communities than by trying to micromanage – and thus delay – network deployment. One better way, for example, might be to issue broadband vouchers or to encourage local non-profits to help pay for access.

This isn’t an either-or problem. Cities still maintain access to public rights of way. But one thing’s for sure. Private firms will be the primary builders of next generation networks. Overwhelmingly so. And faster deployment of wireless networks is good for the public.

This year’s Nobel for economics is a technology prize!

On Tuesday, the Royal Swedish Academy awarded the 2018 Nobel Prize in economic sciences to two American economists, William Nordhaus of Yale University and Paul Romer of New York University’s Stern School of Business. Romer is well-known for his work on innovation, and although the committee focused on Nordhaus’ research on climate change, this year’s prize is really all about technology and its central role in economic growth.

Paul Romer, who with William Nordhaus received the 2018 Nobel Prize in Economics, speaks at the New York University (NYU) Stern School of Business in New York City, October 8, 2018 – via REUTERS

Romer’s 1990 paper “Endogenous technological change” is one of the most famous and cited of the past several decades. Until then, the foundational theory of economic growth was Robert Solow’s model. It said growth was the result of varied quantities of capital and labor, which we could control, and a vague factor known as the Residual, which included scientific knowledge and technology. The Residual exposed a big limitation of the Solow model. Capital and labor were supposedly the heart of the model, and yet technology accounted for the vast bulk of growth — something like 85 percent, compared to the relatively small contributions of capital and labor. Furthermore, technology was an “exogenous” factor (outside our control) which didn’t seem to explain the real world. If technology was a free-floating ever-present factor, equally available across the world, why did some nations or regions do far better or worse than others? (more…)

Indiana, center of the 5G wireless world (at least for today)

About 18 months ago, wireless small cells started popping up all around Indianapolis. The one pictured above is about a half-mile from my house. In addition to these suburban versions, built by one large mobile carrier, a different mobile carrier built a network of 83 small cells in downtown Indy. These small cells are a key architectural facet of the next generation of wireless broadband, known as 5G, and over the next few years we’ll build hundreds of thousands of them across the country. This “densification” of mobile networks will expand coverage and massively boost speeds, responsiveness, and reliability. Our smartphones will of course benefit, but so will a whole range of other new devices and applications.

Building hundreds of thousands of these cells, however, will require lots of investment. A common estimate is $275 billion for the U.S. It will also require the cooperation of states and localities to speed the permitting to place these cells on lampposts, buildings, utility poles, and other rights of way. And this is where Indiana has led the way, with a decade’s worth of pro-broadband policy and, more recently, legislation that’s already encouraged the deployment of more than 1,000 small cells across the state.

Today, Brendan Carr, one of five commissioners of the Federal Communications Commission, visited Indiana to highlight our state’s early successes – and to lay out the next steps in the FCC’s program to expand 5G as quickly as possible. Carr described the key components of his plan, to be voted on at the Commission’s September 25 meeting. The prospective Order:

  1. Implements long-standing federal law that bars municipal rules that have the effect of prohibiting deployment of wireless service
  2. Allows municipalities to charge fees for reviewing small cell deployments when such fees are limited to recovering the municipalities’ costs, and provides guidance on specific fee levels that would comply with this standard
  3. Requires municipalities to approve or disapprove applications to attach small cells to existing structures within 60 days and applications to build new small cell poles within 90 days
  4. Places modest guardrails on other municipal rules that may prohibit service while reaffirming localities’ traditional roles in, for example, reasonable aesthetic reviews

Carr emphasized that this new framework, which will bar excessive fees, will help small towns and communities better compete for infrastructure and capital. We know that wireless firms have to build networks in large “must have” markets such as New York and San Francisco, where millions of Americans live and work. High fees and onerous permitting obstacles, however, are particularly hard on smaller communities – often discouraging investment in these non-urban geographies. This new framework, therefore, is yet another important component of closing the “digital divide.”

Here’s video of Carr’s talk at the Statehouse.

Energy Market of 2030: The End of Carbon Fuels?

See our contribution, with 15 others, to an International Economy symposium looking ahead to the energy market of 2030: The End of Carbon Fuels? Here was our contribution:

The dramatic reduction in U.S. carbon dioxide (CO2) emissions over the last decade is, paradoxically, the result of the massively increased use of a fossil fuel—natural gas. The shale technology revolution produced so much low-cost natural gas, and replaced so much coal, that U.S. emissions from electricity generation have fallen to levels not seen since the late 1980s.

Over time, electric vehicles—and later, autonomous ones—could reduce the need for oil. But natural gas will only rise in importance as the chief generator of inexpensive and reliable electricity.

The Energy Information Administration projects that fossil fuels will still represent 81 percent of total energy consumption in 2030. Natural gas, EIA estimates, will be the largest source of electricity, generating between 50 percent and 100 percent more than renewables.

Sure, but don’t technology revolutions often surprise even the smartest prognosticators? Renewables have indeed been growing from a tiny base, and some believe solar power is poised for miraculous gains.

Despite real advances in solar power and battery storage, however, these technologies don’t follow a Moore’s law path. Solar will grow, but we won’t solve solar’s (nor wind’s) fundamental intermittency and thus unreliability challenges by 2030. Nor can we avoid their voracious appetite for the earth’s surface, a fundamental scarcity which environmentalists and conservationists of all stripes should hope to preserve. Amazon’s Jeff Bezos even dreams of a day when we move much heavy industry into space to preserve the earth’s surface for human enjoyment.

But shouldn’t we pay extra in land area (and dollars) today to avoid CO2’s climate effects tomorrow? Fear not. The latest estimates of the climate’s CO2 sensitivity suggest any warming over the next century will be just half of previous estimates and, therefore, a net benefit to humanity and the earth. Satellites show us that CO2 greens the planet.

Economic growth is the most humane policy today, and it opens up frontiers of innovation, including new energy technologies. Premature anti-CO2 policies can actually boost CO2 emissions, as happened in Germany, where ill-advised wind and solar mandates (and also nuclear decommissionings) so decimated the energy grid that the nation had to quickly build new coal plants. New nuclear technologies are technologically superior to solar and wind but remain irrationally unpopular politically. Emitting more CO2 today may thus accelerate the date when economical, non-CO2 emitting technologies generate most of our power.

Statement on “Restoring Internet Freedom”

Lots of people are asking what I think about today’s FCC vote to roll back the 2015 Title telephone regulations for the Internet, and restore the Internet as an “information service.” So here’s a summary of my view:

Restoring Internet Freedom - statement - Swanson - 12.14.17

The $12-million iPhone

Several years ago, I had a bit of fun estimating how much an iPhone would have cost to make in the 1990s. The impetus was a story making the rounds on the web. A journalist had found a full-page newspaper ad from RadioShack dating back to 1991. He was rightly amazed that all 13 of the advertised electronic gadgets — computer, camcorder, answering machine, cordless phone, etc. — were now integrated into a single iPhone. The cost of those 13 gadgets, moreover, summed to more than $3,000. Wow, he enthused, most of us now hold $3,000 worth of electronics in the palm of our hand.

I saluted the writer’s general thrust but noted that he had wildly underestimated the true worth of our modern handheld computers. In fact, the computing power, data storage capacity, and communications bandwidth of an iPhone in 2014 would have cost at least $3 million back in 1991. He had underestimated the pace of advance by three orders of magnitude (or a factor of 1,000).

Well, in a recent podcast, our old friend Richard Bennett of High Tech Forum brought up the $3 million iPhone 5 from 2014, so I decided to update the estimate. For the new analysis, I applied the same method to my own iPhone 7, purchased in the fall of 2016 — 25 years after the 1991 RadioShack ad. continue reading . . .

Why productivity slowed . . . and why it’s about to soar.

I enjoyed discussing technology’s impact on growth and employment with David Beckworth and Michael Mandel on David’s Macro Musings podcast.

Full speed ahead on the internet

Here’s a brief statement on today’s action at the Federal Communications Commission, where the agency will begin a rule-making to reverse Title II regulation of the Internet and ask how best to protect its freedom and openness.

The Internet has always been open and free, and the successful results were clear for all to see. The imposition of Title II regulation on the Internet in 2015 was unnecessary, illegal, and foolish. Title II was a speed bump that, if allowed to remain, could have grown into a giant road-block to Internet innovation. Fortunately, Chairman Ajit Pai and the FCC today begin the process of returning to the simple rules that for decades fostered Internet investment and entrepreneurship and led to the historically successful digital economy.

The next waves of Internet innovation will bring the amazing power of the digital economy to the physical economy, promising widespread economic benefits. If we want to take the next step, to encourage infrastructure investment and innovation for decades to come, Congress could codify a pro-innovation, pro-consumer approach that would keep the Internet free and open without harmful bureaucratic control.

– Bret Swanson

Robots on TV

See brief interview on Fox Business this morning discussing our “Robots Will Save the Economy” op-ed from The Wall Street Journal.

Robots (and all kinds of info-tech) Will Save the Economy

Screen Shot 2017-05-15 at 9.57.05 AM

See our commentary, with Michael Mandel, in today’s Wall Street Journal: Robots Will Save the Economy.

Ajit Pai’s Welcome Return to Internet Innovation


FCC Chairman Ajit Pai gives speech announcing new approach to Internet regulation, in Washington, D.C., April 26, 2017.

Here’s our latest in Forbes . . . Ajit Pai’s Welcome Return to Internet Innovation:

“Yesterday, Ajit Pai, the new Chairman of the Federal Communications Commission, announced a roll back of the Obama administration’s aggressive regulatory approach to the Internet, adopted in early 2015. This afternoon, Pai will release the text of the proposed rule-making, which will launch several months of public comment.

“In yesterday’s speech, Pai emphasized the Internet’s historic success, based on a bipartisan approach adopted in the Clinton administration, which elevated innovation over regulation by consciously rejecting the old telephone rules for the emerging digital economy:

Under this framework, a free and open Internet flourished. Under this framework, America’s Internet economy produced the world’s most successful online companies: Google, Facebook, and Netflix, just to name a few. Under this framework, the private sector invested about $1.5 trillion to build the networks that gave people high-speed access to the Internet. And under this framework, consumers benefited from unparalleled innovation. But two years ago, the federal government’s approach suddenly changed.

continue reading . . .

Dawn? Or doom? I vote dawn.

Here is video of a presentation from last October’s Dawn or Doom technology conference at Purdue University, where I previewed the Coming Productivity Boom research.

« Previous PageNext Page »