If laissez-faire politics is dead, how big a role should the state have? The economic crisis offers us a chance to rethink the role of government. Ten leading thinkers from Britain and Germany mull the question in a Comment is free special in association with the Global Policy Institute at London Metropolitan University and the Friedrich Ebert Foundation London Office

With every crisis comes the need for change, and the current financial and economic crisis is no exception to this rule. Even though it will take much more time for a new, remodelled international system to emerge, some striking changes resulting from the current predicament are beginning to emerge.

Probably the biggest transformation so far has been the comeback of the state. All over the world, government action on an unprecedented scale was needed to prevent the financial system from total collapse. And now, at the beginning of a global recession, governments are again taking the initiative and trying to soften the coming economic blows with massive financial stimulus packages.

This comeback came as a surprise to many. Given the relative decline of state influence since the ascendancy of the neo-liberal paradigm, few would have predicted such a return of the state only a few months ago. But the current crisis has shown that only the state – or systems of democratic governance, to be more precise – has the ability and legitimacy to act effectively in times of turmoil.

So a crucial question arises: What is the state's role in the future? Which functions should it fulfil and which should be left to market forces? Where is the new line between the private and the public? There are many opinions on this question, but the current discussions have an important shortcoming: they suffer from a lack of structure. With this series of articles, we will attempt to frame the debate.

This series is effectively a brainstorm of ten eminent progressive public intellectuals from the UK and Germany. Each author was asked to answer the same question: What future role do you see for the state? The combination of recurring themes and new perspectives brought up by each author provide a fruitful basis for deeper investigation and discussion. This series is certainly not an exhaustive account of all relevant ideas, but we hope to provide a valuable collection of fresh thinking.

Of course, the issue can't be covered comprehensively on a merely British-German basis. But you have to start somewhere. We would therefore like to invite other voices from Europe and beyond to join us in helping to shape the central debate.

This article series is dedicated to the memory of our colleague and friend Professor Detlev Albers, whose untimely death earlier this year deeply saddened us. His contributions to this discussion are dearly missed.

One of the most amazing things to me in the recent electoral process was the complete absence of serious discussion by either camp of anything relating to Latin America.

While Venezuela, Nicaragua, Bolivia and Ecuador have formed a new anti-U.S. and pro-Iranian axis, perhaps the most dangerous developments are taking place directly on our southern border. It is to the point where the Mexican government, despite commendable efforts by the Calderon administration, is barely hanging on.

I say this with deep sadness. I lived through the similar Colombian experience where the narco-violence shattered the country, almost took over the state, and claimed the lives of many of my friends, including journalists I worked with. That is part of the reason I cannot understand the lack of attention to the crisis.

We are looking at the creation of a series of mini narco-states along our border, and from our border heading south through Central America. Mexico’s own southern border with Guatemala is now awash in cartel violence, and much of Guatemala is no longer under state control. El Salvador is a money laundering sanctuary, as is Panama, and Nicaragua, along with Venezuela, have become black holes where an increasing amount of cocaine transits.

This is a clear and present danger not only to the United States, which will (and already has) suffer from the spillover of violence and border security-but to the Mexican state as well.

Senior police and government officials are assassinated with such frequency their deaths seldom make the news here at all. Some are killed because they are trying to do the right thing, some because they are on the losing side of an intra-cartel battle, but it is shredding the authority of the state.

Illegal immigration of non-Latin Americans, including Iranians, Chinese and Turks, is on the rise. Why does this matter? Because the cartels in their many different iterations and factions, thrive on instability, and their ability to move anything (drugs, people, weapons) across our border. Iranians and Turks arriving this way are not here to see the Empire State Building or the Grand Canyon.

The collapse of Mexico into the hands of drug traffickers is exacerbated by the global financial crisis. Remittances from the United States have dropped sharply, and Mexican migrants are returning home in droves. This has its positive side, but right now it means that more and more people will turn to drug trafficking to make a living, strengthening the narco side of the conflict.

This crisis has to be (another) priority for the Obama administration. The Bush administration seemed content to work on the margins and ignore the realities on the ground as long as possible. Obama will not have that luxury.

jueves, 11 de diciembre de 2008

Behind the debate over remaking US financial policy will be a debate over who's to blame. It's crucial to get the history right, writes a Nobel-laureate economist, identifying five key mistakes - under Reagan, Clinton and Bush II - and one national delusion.

There will come a moment when the most urgent threats posed by the credit crisis have eased and the larger task before us will be to chart a direction for the economic steps ahead. This will be a dangerous moment. Behind the debates over future policy is a debate over history - a debate over the causes of our current situation. The battle for the past will determine the battle for the present. So it's crucial to get the history straight.

What were the critical decisions that led to the crisis? Mistakes were made at every fork in the road - we had what engineers call a "system failure," when not a single decision but a cascade of decisions produce a tragic result. Let's look at five key moments.

No. 1: Firing the Chairman

In 1987 the Reagan administration decided to remove Paul Volcker as chairman of the Federal Reserve Board and appoint Alan Greenspan in his place. Volcker had done what central bankers are supposed to do. On his watch, inflation had been brought down from more than 11 percent to under 4 percent. In the world of central banking, that should have earned him a grade of A+++ and assured his re-appointment. But Volcker also understood that financial markets need to be regulated. Reagan wanted someone who did not believe any such thing, and he found him in a devotee of the objectivist philosopher and free-market zealot Ayn Rand.

Greenspan played a double role. The Fed controls the money spigot, and in the early years of this decade, he turned it on full force. But the Fed is also a regulator. If you appoint an anti-regulator as your enforcer, you know what kind of enforcement you'll get. A flood of liquidity combined with the failed levees of regulation proved disastrous.

Greenspan presided over not one but two financial bubbles. After the high-tech bubble popped, in 2000-2001, he helped inflate the housing bubble. The first responsibility of a central bank should be to maintain the stability of the financial system. If banks lend on the basis of artificially high asset prices, the result can be a meltdown - as we are seeing now, and as Greenspan should have known. He had many of the tools he needed to cope with the situation. To deal with the high-tech bubble, he could have increased margin requirements (the amount of cash people need to put down to buy stock). To deflate the housing bubble, he could have curbed predatory lending to low-income households and prohibited other insidious practices (the no-documentation - or "liar" - loans, the interest-only loans, and so on). This would have gone a long way toward protecting us. If he didn't have the tools, he could have gone to Congress and asked for them.

Of course, the current problems with our financial system are not solely the result of bad lending. The banks have made mega-bets with one another through complicated instruments such as derivatives, credit-default swaps, and so forth. With these, one party pays another if certain events happen - for instance, if Bear Stearns goes bankrupt, or if the dollar soars. These instruments were originally created to help manage risk - but they can also be used to gamble. Thus, if you felt confident that the dollar was going to fall, you could make a big bet accordingly, and if the dollar indeed fell, your profits would soar. The problem is that, with this complicated intertwining of bets of great magnitude, no one could be sure of the financial position of anyone else - or even of one's own position. Not surprisingly, the credit markets froze.

Here too Greenspan played a role. When I was chairman of the Council of Economic Advisers, during the Clinton administration, I served on a committee of all the major federal financial regulators, a group that included Greenspan and Treasury Secretary Robert Rubin. Even then, it was clear that derivatives posed a danger. We didn't put it as memorably as Warren Buffett - who saw derivatives as "financial weapons of mass destruction" - but we took his point. And yet, for all the risk, the deregulators in charge of the financial system - at the Fed, at the Securities and Exchange Commission, and elsewhere - decided to do nothing, worried that any action might interfere with "innovation" in the financial system. But innovation, like "change," has no inherent value. It can be bad (the "liar" loans are a good example) as well as good.

No. 2: Tearing Down the Walls

The deregulation philosophy would pay unwelcome dividends for years to come. In November 1999, Congress repealed the Glass-Steagall Act - the culmination of a $300 million lobbying effort by the banking and financial-services industries, and spearheaded in Congress by Senator Phil Gramm. Glass-Steagall had long separated commercial banks (which lend money) and investment banks (which organize the sale of bonds and equities); it had been enacted in the aftermath of the Great Depression and was meant to curb the excesses of that era, including grave conflicts of interest. For instance, without separation, if a company whose shares had been issued by an investment bank, with its strong endorsement, got into trouble, wouldn't its commercial arm, if it had one, feel pressure to lend it money, perhaps unwisely? An ensuing spiral of bad judgment is not hard to foresee. I had opposed repeal of Glass-Steagall. The proponents said, in effect, Trust us: we will create Chinese walls to make sure that the problems of the past do not recur. As an economist, I certainly possessed a healthy degree of trust, trust in the power of economic incentives to bend human behavior toward self-interest - toward short-term self-interest, at any rate, rather than Tocqueville's "self interest rightly understood."

The most important consequence of the repeal of Glass-Steagall was indirect - it lay in the way repeal changed an entire culture. Commercial banks are not supposed to be high-risk ventures; they are supposed to manage other people's money very conservatively. It is with this understanding that the government agrees to pick up the tab should they fail. Investment banks, on the other hand, have traditionally managed rich people's money - people who can take bigger risks in order to get bigger returns. When repeal of Glass-Steagall brought investment and commercial banks together, the investment-bank culture came out on top. There was a demand for the kind of high returns that could be obtained only through high leverage and big risktaking.

There were other important steps down the deregulatory path. One was the decision in April 2004 by the Securities and Exchange Commission, at a meeting attended by virtually no one and largely overlooked at the time, to allow big investment banks to increase their debt-to-capital ratio (from 12:1 to 30:1, or higher) so that they could buy more mortgage-backed securities, inflating the housing bubble in the process. In agreeing to this measure, the S.E.C. argued for the virtues of self-regulation: the peculiar notion that banks can effectively police themselves. Self-regulation is preposterous, as even Alan Greenspan now concedes, and as a practical matter it can't, in any case, identify systemic risks - the kinds of risks that arise when, for instance, the models used by each of the banks to manage their portfolios tell all the banks to sell some security all at once.

As we stripped back the old regulations, we did nothing to address the new challenges posed by 21st-century markets. The most important challenge was that posed by derivatives. In 1998 the head of the Commodity Futures Trading Commission, Brooksley Born, had called for such regulation - a concern that took on urgency after the Fed, in that same year, engineered the bailout of Long-Term Capital Management, a hedge fund whose trillion-dollar-plus failure threatened global financial markets. But Secretary of the Treasury Robert Rubin, his deputy, Larry Summers, and Greenspan were adamant - and successful - in their opposition. Nothing was done.

No. 3: Applying the Leeches

Then along came the Bush tax cuts, enacted first on June 7, 2001, with a follow-on installment two years later. The president and his advisers seemed to believe that tax cuts, especially for upper-income Americans and corporations, were a cure-all for any economic disease - the modern-day equivalent of leeches. The tax cuts played a pivotal role in shaping the background conditions of the current crisis. Because they did very little to stimulate the economy, real stimulation was left to the Fed, which took up the task with unprecedented low-interest rates and liquidity. The war in Iraq made matters worse, because it led to soaring oil prices. With America so dependent on oil imports, we had to spend several hundred billion more to purchase oil - money that otherwise would have been spent on American goods. Normally this would have led to an economic slowdown, as it had in the 1970s. But the Fed met the challenge in the most myopic way imaginable. The flood of liquidity made money readily available in mortgage markets, even to those who would normally not be able to borrow. And, yes, this succeeded in forestalling an economic downturn; America's household saving rate plummeted to zero. But it should have been clear that we were living on borrowed money and borrowed time.

The cut in the tax rate on capital gains contributed to the crisis in another way. It was a decision that turned on values: those who speculated (read: gambled) and won were taxed more lightly than wage earners who simply worked hard. But more than that, the decision encouraged leveraging, because interest was tax-deductible. If, for instance, you borrowed a million to buy a home or took a $100,000 home-equity loan to buy stock, the interest would be fully deductible every year. Any capital gains you made were taxed lightly - and at some possibly remote day in the future. The Bush administration was providing an open invitation to excessive borrowing and lending - not that American consumers needed any more encouragement.

No. 4: Faking the Numbers

Meanwhile, on July 30, 2002, in the wake of a series of major scandals - notably the collapse of WorldCom and Enron - Congress passed the Sarbanes-Oxley Act. The scandals had involved every major American accounting firm, most of our banks, and some of our premier companies, and made it clear that we had serious problems with our accounting system. Accounting is a sleep-inducing topic for most people, but if you can't have faith in a company's numbers, then you can't have faith in anything about a company at all. Unfortunately, in the negotiations over what became Sarbanes-Oxley a decision was made not to deal with what many, including the respected former head of the S.E.C. Arthur Levitt, believed to be a fundamental underlying problem: stock options. Stock options have been defended as providing healthy incentives toward good management, but in fact they are "incentive pay" in name only. If a company does well, the C.E.O. gets great rewards in the form of stock options; if a company does poorly, the compensation is almost as substantial but is bestowed in other ways. This is bad enough. But a collateral problem with stock options is that they provide incentives for bad accounting: top management has every incentive to provide distorted information in order to pump up share prices.

The incentive structure of the rating agencies also proved perverse. Agencies such as Moody's and Standard & Poor's are paid by the very people they are supposed to grade. As a result, they've had every reason to give companies high ratings, in a financial version of what college professors know as grade inflation. The rating agencies, like the investment banks that were paying them, believed in financial alchemy - that F-rated toxic mortgages could be converted into products that were safe enough to be held by commercial banks and pension funds. We had seen this same failure of the rating agencies during the East Asia crisis of the 1990s: high ratings facilitated a rush of money into the region, and then a sudden reversal in the ratings brought devastation. But the financial overseers paid no attention.

No. 5: Letting It Bleed

The final turning point came with the passage of a bailout package on October 3, 2008 - that is, with the administration's response to the crisis itself. We will be feeling the consequences for years to come. Both the administration and the Fed had long been driven by wishful thinking, hoping that the bad news was just a blip, and that a return to growth was just around the corner. As America's banks faced collapse, the administration veered from one course of action to another. Some institutions (Bear Stearns, A.I.G., Fannie Mae, Freddie Mac) were bailed out. Lehman Brothers was not. Some shareholders got something back. Others did not.

The original proposal by Treasury Secretary Henry Paulson, a three-page document that would have provided $700 billion for the secretary to spend at his sole discretion, without oversight or judicial review, was an act of extraordinary arrogance. He sold the program as necessary to restore confidence. But it didn't address the underlying reasons for the loss of confidence. The banks had made too many bad loans. There were big holes in their balance sheets. No one knew what was truth and what was fiction. The bailout package was like a massive transfusion to a patient suffering from internal bleeding - and nothing was being done about the source of the problem, namely all those foreclosures. Valuable time was wasted as Paulson pushed his own plan, "cash for trash," buying up the bad assets and putting the risk onto American taxpayers. When he finally abandoned it, providing banks with money they needed, he did it in a way that not only cheated America's taxpayers but failed to ensure that the banks would use the money to restart lending. He even allowed the banks to pour out money to their shareholders as taxpayers were pouring money into the banks.

The other problem not addressed involved the looming weaknesses in the economy. The economy had been sustained by excessive borrowing. That game was up. As consumption contracted, exports kept the economy going, but with the dollar strengthening and Europe and the rest of the world declining, it was hard to see how that could continue. Meanwhile, states faced massive drop-offs in revenues - they would have to cut back on expenditures. Without quick action by government, the economy faced a downturn. And even if banks had lent wisely - which they hadn't - the downturn was sure to mean an increase in bad debts, further weakening the struggling financial sector.

The administration talked about confidence building, but what it delivered was actually a confidence trick. If the administration had really wanted to restore confidence in the financial system, it would have begun by addressing the underlying problems - the flawed incentive structures and the inadequate regulatory system.

Was there any single decision which, had it been reversed, would have changed the course of history? Every decision - including decisions not to do something, as many of our bad economic decisions have been - is a consequence of prior decisions, an interlinked web stretching from the distant past into the future. You'll hear some on the right point to certain actions by the government itself - such as the Community Reinvestment Act, which requires banks to make mortgage money available in low-income neighborhoods. (Defaults on C.R.A. lending were actually much lower than on other lending.) There has been much finger-pointing at Fannie Mae and Freddie Mac, the two huge mortgage lenders, which were originally government-owned. But in fact they came late to the subprime game, and their problem was similar to that of the private sector: their C.E.O.'s had the same perverse incentive to indulge in gambling.

The truth is most of the individual mistakes boil down to just one: a belief that markets are self-adjusting and that the role of government should be minimal. Looking back at that belief during hearings this fall on Capitol Hill, Alan Greenspan said out loud, "I have found a flaw." Congressman Henry Waxman pushed him, responding, "In other words, you found that your view of the world, your ideology, was not right; it was not working." "Absolutely, precisely," Greenspan said. The embrace by America - and much of the rest of the world - of this flawed economic philosophy made it inevitable that we would eventually arrive at the place we are today.