Robert Gordon of Northwestern University is a distinguished economist whose work in macroeconomics and studies of long-term economic growth have properly earned him high regard. So his recent exercise in speculative future history, which asks whether economic growth in the U.S. has come to an end, has attracted much favorable attention. But a basic flaw in Gordon's argument is immediately apparent and becomes glaringly so on closer examination.
Gordon distinguishes three industrial revolutions that have driven economic growth and improved living standards since the 18th century: industrial revolution No. 1 (steam, railroads), whose defining inventions date from between 1750 and 1830; industrial revolution No. 2 (electricity, internal combustion engine, running water, indoor toilets, communications, entertainment, chemicals, petroleum), whose defining inventions date from 1870 to 1900; and industrial revolution No. 3 (computers, the Internet, cell phones), dating from 1960. The core of his article contrasts the transformational impact of industrial revolution No. 1 and especially industrial revolution No. 2, based on per capita gross domestic product and quality of life, with the relatively trivial consequences of industrial revolution No. 3.
The vulnerability of Gordon's argument is his shortened time horizon for industrial revolution No. 3. Consider the following four sentences in his paper:
• "Both the first two revolutions required about 100 years for their full effects to percolate through the economy."
• "At a minimum, it took 150 years for Industrial Revolution No. 1 to have its full range of effects."
• "The inventions of Industrial Revolution No. 2 were so important … that they took a full 100 years to have their main effect."
• "The productivity impact of Industrial Revolution No. 3 evaporated after only eight years, compared to the 81 years (1891-1972) required for the benefits of Industrial Revolution No. 2 to have their full impact."
The last sentence is crucial. Gordon cuts off industrial revolution No. 3 circa 2005 — that is, 45 years after its onset but less than half the time allowed for industrial revolution No. 1 and industrial revolution No. 2 to run their respective courses. To take a salient example from the prior industrial revolutions, this is making calculations as if the impact of the railroads on the U.S. economy were measured as of 1873, 45 years after construction began on the United States' first line, the Baltimore & Ohio.
In 1873, the retailer Montgomery Ward was just a year old, and the first Sears Roebuck catalogue was still 20 years away. Together, these companies invented mail order as the "killer app" of the railroad age, creating a continental market for consumer goods, with all of the economies of scale that followed.
Or take another example. What if we stopped measuring the economic impact of electrification only 45 years after the first generating plant, Thomas Edison's Pearl Street station, became operational in 1882? At that point, the United States' manufacturing industries were just discovering the benefits of flexible (re)configuration, which enabled distribution of generated electric power, while the U.S. home-appliance industry was in its infancy.
Gordon asserts that "the era of computers replacing human labor was largely over" during the past decade. Innovation in information and communications technology focused first on electronic commerce, "itself largely completed by 2005." After that, "labor-saving innovation" took a back seat to "a succession of entertainment and communication devices that do the same things as we could do before, but now in smaller and more convenient packages."
In thus truncating and trivializing the ongoing information and communications technology revolution, Gordon misses two fundamental processes. First, and most evident, the rise of e-commerce is far from over. In the U.S., the most advanced country in this respect, e-commerce just reached 10 percent of total retail spending and continues to grow at double-digit rates through the slow recovery from the Great Recession. At the same time, something much more significant is happening beneath the economic surface.
From the early days of the computer revolution, researchers and popularizers envisioned "artificial intelligence" as the ultimate killer app (literally, in the case of the computer HAL in the film "2001: A Space Odyssey"). Decades of frustration are now yielding to success: The application of sophisticated statistical techniques to the accelerating accumulation of unprecedented quantities of so-called Big Data, which the Internet simultaneously generates and captures. For proof of artificial intelligent "life," consider the "virtual assistants" populating smartphones, beginning — but just beginning — with Apple's Siri.
The next wave of industrial revolution No. 3's consequences can be discerned in systems that capture the intention behind a search request and couple it to a completed transaction and in the application of predictive analytics to industrial supply chains and service provisioning. If these innovations' prospective impact on productivity and GDP growth cannot be quantified, well, that is the way it is with future history.
Flawed as Gordon's reading of industrial revolution No. 3 is, his concluding concern for the future of innovation in the U.S. is well-founded. Of the assorted "headwinds" that he cites, the enormous increase in inequality and the plateau reached in educational attainment represent threats that differentiate the U.S. position from that of other advanced countries. But Gordon ignores the greatest threat to continued U.S. global leadership in innovation: the delegitimation over the past generation of the state's role as an economic actor.
The state has become central to the innovation economy's dynamics. It funds the upstream research that generates discovery and invention, supports the creation of new networks, from canals to the Internet, and serves as a creative customer for innovative products, as it did during the computer revolution's nascence.
The state must also preserve continuity in the market economy when the speculative bubble that has funded its transformation bursts. Here, too, Gordon disappoints. The only reference to government in his paper is not even to this stabilizing role. Rather, it is to the asserted headwind of the debt incurred to prevent the global financial crisis of 2008 from triggering a second great depression.
William Janeway, a managing director and senior adviser at the private equity firm Warburg Pincus, is a visiting lecturer in economics at Cambridge University and author of "Doing Capitalism in the Innovation Economy: Markets, Speculation and the State." © Project Syndicate
Related articles:
A Message from The Moscow Times:
Dear readers,
We are facing unprecedented challenges. Russia's Prosecutor General's Office has designated The Moscow Times as an "undesirable" organization, criminalizing our work and putting our staff at risk of prosecution. This follows our earlier unjust labeling as a "foreign agent."
These actions are direct attempts to silence independent journalism in Russia. The authorities claim our work "discredits the decisions of the Russian leadership." We see things differently: we strive to provide accurate, unbiased reporting on Russia.
We, the journalists of The Moscow Times, refuse to be silenced. But to continue our work, we need your help.
Your support, no matter how small, makes a world of difference. If you can, please support us monthly starting from just $2. It's quick to set up, and every contribution makes a significant impact.
By supporting The Moscow Times, you're defending open, independent journalism in the face of repression. Thank you for standing with us.
Remind me later.