Skip to content

[RsR] [R-SIG-Finance] Outliers in the market model that's used to estimate `beta' of a stock

2 messages · m@rkieeds m@iii@g oii verizo@@@et, Matias Salibian-Barrera

#
Hi: i don't know if you read "fooled by randomness" by Nassim Taleb ( 
spelling )  but he essentially says using very non statistical arguments 
but
strong nevertheless. ( it's not a stat or a quant finance book )  that 
outliers in finance are not modellable and don't claim that you can 
model
them because you'd be lying. In fact, he would say that a model works 
until it doesn't.

Anyway, it's an interesting book that sort of indirectly talks ( for a 
little too long actually. you can get what's he saying in the first 50 
pages and
  it's about 200 pages )  about your comment below so I figured I would 
just mention it in case you were interested.
On Thu, Sep 18, 2008 at 11:36 AM, Ajay Shah wrote:

            
#
Haven't read "Fooled by randomness", but did start reading Black Swan, 
and although in general I like provocative books that challenge my 
points of view, I found his main thesis to be too short to warrant so 
many words... I took it that his main argument was with those who 
misinterpret and misuse statistics (particularly when they do it for 
their own benefit), not with statistics itself, which is always based on 
assumptions etc.
Which is a fair statement, that also applies to science in general, 
"theories work until they are proved wrong", and the whole 
"falsifiability" argument (cf. Popper vs. Kuhn vs. Feyerabend vs...). I 
believe robust statistics can help you determine when your model 
(theory) has stopped to work.

In any case, with respect to the old "data cleaning versus robust 
estimators" discussion, I would point the interest reader to the first 
chapter of Maronna, Martin and Yohai's book 
(http://books.google.com/books?id=YD--AAAACAAJ&dq=martin+maronna+yohai), 
and for some more specific inference implications, to the first chapter 
of my PhD dissertation. Essentially, a couple of main issues are: (a) 
detecting outliers using non-robust estimators does not work well in 
general (but even if / when it does, see my next point); (b) if you 
remove (or alter) observations, all subsequent probabilistic statements 
(p-values, standard errors, etc) are all conditional on the very 
non-linear cleaning operation you did, and thus both wrong at face 
value, and not easy to correct. Robust estimators incorporate the 
down-weighting and its effect on the corresponding inference at once, 
and are thus, IMHO, to be preferred.

Matias
markleeds at verizon.net wrote: