Hacker Timesnew | past | comments | ask | show | jobs | submitlogin

> It is a low-probability, high-impact event that, because of its rarity and the instability of the environment, cannot be scientifically evaluated in terms of risk and return.

I stopped reading here. The only objective evaluation of economic matters is in scientific, risk-and-return terms. Anything else is quackery, fear-mongering, or worse.

A great deal of evil has been done in the last decade or so in the name of protecting us against horrible things. We are told that much good has come of it, because nothing horrible has happened. While I'm sure this is partly true, I wouldn't mind trading a few more horrible things for a little more autonomy.



"The only objective evaluation of economic matters is in scientific, risk-and-return terms."

That's not the problem; I suggest you read up on his theories. His point is, to put it more mathematically, that using inappropriate statistical models to try to model future events can lead to catastrophe. In particular, assuming something is a normal Gaussian process when in fact it's something that's almost a normal Gaussian but has a slight propensity towards total disasters that occur more often (and much stronger) than the Gaussian distribution would imply can result in you creating a system that is terribly brittle when those "black swans" occur.

"Scientifically" applying the Gaussian distribution or other advanced mathematical constructs when they manifestly do not apply is the real quackery, if Taleb is right. And it's hard to look at the evidence and conclude he's wrong here; it's simply objective fact that many of the interesting distributions are in fact prone to "black swans", there's a statistical test that shows it (kurtosis) quite clearly. It's really more a question of what we are going to do about his point than whether or not he's correct about the distribution. (There's no guarantee that his prescription is correct, but it's probably a better idea than listening to the "experts" who advise you to mortgage to the hilt because the worst case scenario can't possibly happen and you are 99.9% percently likely to be fine. You know, that worst case scenario that just happened and we're still living through. Yeah, that scenario that can't possibly happen.)


I agree about the statistical issues he highlights, but his anti expert stance is kind of odd. I mean we listen to this guy because he used to be a wall street trader. His books were published because he was accepted as some kind of an expert himself. So is he wrong because he's an expert?

He says: "We replaced the heuristics of the elders with arrogant (and incompetent) beliefs, breaking, in the name of science, the chain of knowledge."

In europe we had fascisim and stalinism within the past 100 years. That's three or four generations. Before the dawn of liberalism, capitalism, enlightenment and the industrial revolution we had 1000 years of depression caused by religious fanatics. We had to break that chain of "knowledge".

The heuristics of the elders are little more than mature, cultivated cliches of what past experts and ideologues told them. Taleb is correct, it's incredibly difficult to make predictions in a complex system. However, I don't see how a blanket condemnation of science and a blanket approval of crude traditions helps us to become better at that.


Black Swan theory isn't about denying science; it's about recognizing the limits of science.

Taleb's position (and argument in this paper) is that we need to recognize instances where we don't know the answer instead of making decisions based on assumptions that we do.


You can't discount the apocalypse. You just can't. And this pan-phobic concept "black swan" is just various degrees of unknowable apocalypses that we are supposed to account for based upon their very unaccountableness. I don't buy it. Read the tea leaves first, then tell me about swan feathers.

The housing bubble was BLATANT. No black swans involved. I remember talking about it with friends in 2005. No big deal, really.


"The housing bubble was BLATANT. No black swans involved."

Yes, but what was not blatant was that the housing bubble popping would kill Lehman Brothers, cause runs on banks and almost bring down the whole world economy.

Taleb's point is that you can't go around behaving like any system is, to borrow a phrase "too big to fail" - or rather too cleverly thought out/regulated to fail. To some extent this is simple engineering. I can test the resilience of my cluster to network failure by pulling out an ethernet cable and see what happens - maybe there's a small hiccup while systems fall back onto local resources, maybe nothing fails over properly and I am hosed for 12 hours. As I understand Taleb, his argument is that the economic powers (bankers, traders, politicans) spend too much time trying to figure out how to prevent network failure, rather than making sure that when the network does go out, it doesn't take your whole mission down with it. Because, no matter how well you think you have engineered the system, the chance of your network going out somehow, sometime is non-zero.

I do agree with the posters that Taleb overstates the cleverness of his own insight; but unfortunately, in his field, it seems that people really were too dumb/greedy to grasp this. Irrespective of what you think of his interpretation of events or the validity of his suggestions (and I too question them), I think as systems engineers we can appreciate the risk of designing political and economic systems under the assumption that they can't fail.


the problem isn't that people were too dumb/greedy to notice. it's that they weren't punished for not noticing it. you'd be amazed what a little incentive does to people's willingness to self educate.


If it really was that blatant, why did it happen? Doesn't the fact that it burst by definition mean that most people did not recognize it?

But Taleb makes a different point: Discounting the apocalypse is exactly what he's arguing against, because this is essentially what LTCM et al were doing with their models. They calculated probabilities to make sure that the estimated payoff was greater than the estimated loss, but they got the distributions wrong. What you should try to do is make sure that the apocalypse can't happen, that you aren't exposed so that one outlier will take you down. Because due to the nature of the system (there is only a sample of one of the world economy) you simply can't gather data that can pin down the wings of the distributions to the precision you need.

According to the models used at the time, the Black Monday 1987 was something like a 30-sigma event. If the distributions truly were gaussian, such an event should never happen in the history of the Universe. So you have a choice between thinking that we just had an incredibly bad luck, or the model is wrong. I'd go with the latter.


"If it really was that blatant, why did it happen? Doesn't the fact that it burst by definition mean that most people did not recognize it?"

Some data, some math, and an interpretation you won't hear every day:

http://quickfacts.census.gov/qfd/states/00000.html

http://photos1.blogger.com/photoInclude/x/blogger/2825/754/1...

There are 127 million housing units in the U.S. At the peak of the bubble, 7 million houses/year were sold. A total of roughly 30 million houses were sold in the boom years of 2002-2007, assuming there's no double-counting (people who sold a house multiple times), which is probably a pessimistic assumption. That means that over half of U.S. households took absolutely no part in the bubble - they just sat tight.

Now look at the inventory numbers from the last two years. They go up rather dramatically. This means there were more sellers than buyers, i.e. a large number of people realized that home prices were selling for more than they thought the house was worth, and quickly put their house on the market to take advantage of this. Sure looks like people recognized the bubble to me, it's just that by definition they can't all get out before it bursts.

My interpretation is that most people did recognize the bubble - it's just that they're not the ones that the New York Times writes stories about. They held onto their home and neither bought nor sold, instead just going about their business. Or they sold at the peak, rented for a bit, and are now buying back in at half price and pocketing a few hundred grand. Smart people tend not to brag about how they just made a killing off other people's stupidity. It prevents people from being stupid again, which limits the opportunities for future killings. Besides, it tends to kill the mood at parties.


Many people knew there was a housing bubble, but few predicted it would have such wide-reaching effects. Did you?


I had all my money in the foreign markets, and got out in late 2007 when it spiked down,

http://finance.yahoo.com/q/bc?s=VEIEX&t=5y

...I wouldn't have made any connection to US housing and foreign equities, but I was looking at both, and I knew both were bubbly. I've since lost money on a few stupid trades, so it's not like I'm an expert or anything. But the fundamentals are usually clear to all. Right now, the macros on S&P PEs, long-term treasuries, and high-yield,

http://finance.yahoo.com/q/bc?s=JNK&t=2y

...are clear, and they all scream "avoid". And that's the extent of my trading strategy right now.


"Taleb's position (and argument in this paper) is that we need to recognize instances where we don't know the answer instead of making decisions based on assumptions that we do."

Have you ever worked at an investment bank? I suspect you have not. Wall Street does not run on the scientific method. It runs on greed, fear and politics. The ones who check the assumptions and know the limits of their models don't have the power to make the decisions. This is the problem. And the Taleban apparently haven't yet figured it out. Quoting Janet Tavakoli: "malfeasance, not models, disrupted the global economy, and Taleb still gets that part wrong."

Last but not least: there's not such thing as a "blackswan theory". You can't steal ideas from probability theory and statistics that have been known for centuries, rename them, wrap them up nicely, and call them new or original.


Malfeasance enabled the models, the models enabled malfeasance. Declaring one thing "the culprit" and not the other is to fail to recognize the dynamic nature of the system. Even in a world of perfect angels, Taleb shows that broken models would still bring the system down. Even in a world of perfect models, unscrupulous jerks can bring the system down. It's all too related to pull apart.


So, what is the solution? Destroy the financial industry? Good luck.

Taleb is an anti-intellectual. He believes that all models are wrong and useless. To make it worse, he advocates that no further knowledge can be attained. The only knowledge and truth that is right is the "black swan theory". Come on, gimme a break!!! Aren't we HN'ers smarter than that? Why is Taleb's drivel being upvoted?!


The idea I've gotten from reading Taleb (articles, not books) is not that all models are useless, it's that assuming a normal distribution is often unjustified, and that once-in-a-century events come along quite a bit more often than that. Sure; he invented a catchphrase, not the exponential distribution--but "Black Swan" is more than just a cheap Australian Cabernet if it'll generate more positive awareness of systemic risks.


"You can't steal ideas from probability theory and statistics that have been known for centuries, rename it, wrap it nicely and call it a new, original idea."

Do you have any references to work centuries old that describes what Taleb calls 'Black Swans'?


No need for references. Probability was started by a bunch of bored french aristocrats who wanted to understand risk and uncertainty. Read what they wrote. They're the original thinkers.

Last but not least: in the late 1990s, Lehman Brothers had a trading desk dedicated solely to extremely unlikely events. That was before the so-called "black swan" theory came along. They made money only once. On 9/11.

Taleb's good ideas are not original, and his original ideas are no good.


To be fair, the work of Pascal and the Chevalier de Mere does not give much insight on how a water leak may turn into a mudslide. Whatever you may think of Taleb, his work is more of a popularization of catastrophe theory than it is basic probability. He even says as much in his writings.


I think if you put that statement in context, bin Taleb is saying that such "black swan" events can't be understood in the absence of a suitable behavioral model for the economy -- and furthermore, you don't want to rely on any global behavior models because they ultimately prove limiting. Communism anyone?

There is a lot going on in finance every day that simply cannot be objectively modeled or understood in scientific terms. That's a problem, but we also need to accept that there may not be a globally valid solution. Viewing complexity as the enemy sounds like as good an approach as any.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: