Re: rms of indices

From: John Conover <john@email.johncon.com>
Subject: Re: rms of indices
Date: 3 Jan 2001 09:58:09 -0000



Note that this does not validate the analytical method, (tsinvest.)

What it does validate is being able to synthesize the stocks in a
market, (tsinvestsim,) at least such that the analytical method can
not determine whether the data is empirical or fabricated.

So, the synthesizer, (tsinvestsim,) is used to produce very precise
characteristics of stocks in a market, (which can be verified
theoretically,) and the analytical method, (tsinvest,) tested to see
if it produces the expected theoretical values-to gain confidence in
the analytical method.

Its a long drawn out process, (struggling through the numbers, the
simulation has to run for a minimum of 4 centuries of daily data for a
minimum of 300 stocks-all as dictated by the tsshannoneffective
program,) and involves running corner solutions in a three sided
envelope, (non-volatile, optimal, and volatile, for stocks that are
the same, and different, in all combinations thereof; and then the
same scenario for stocks with persistent and anti-persistent
characteristics.)

The results of the testing process take about 2 weeks of a dedicated
Pentium class machine to produce, and are tabulated on the tests page,
http://www.johncon.com/ntropix/tests.html, where they are compared
against the theoretical numbers, (and referenced in the first
paragraph of the quality assurance page,
http://www.johncon.com/ntropix/QA.html.)

Note that the entire QA and testing process dependeds on the utilities
listed on http://www.johncon.com/ntropix/utilities.html.

        John

BTW, the validation process used is by no means fool proof, (its
circular logic if you look at it.) But what it does do is lend
confidence to the analytical methods used prior to empirical
verification.

If you are a glutton for punishment, try something like this with
MACD, etc.

John Conover writes:
> This is why the second paragraph of the quality assurance page,
> http://www.johncon.com/ntropix/QA.html, works.
>
> Note that all that was done was to extract the analytical data from a
> set of stocks, (three numbers for each stock,) and use those numbers
> in an inverse function, (simulation,) to reconstruct the original
> data-and then analyze the reconstructed data, and compare the original
> analytical data with the reconstructed analytical data.
>
>         John
>
> John Conover writes:
> > And, then if you are a real glutton for punishment, you can write a
> > program that does a point-by-point reconstruction of the indice using
> > the formula:
> >
> >     G = (1 + rms)^P * (1 - rms)^(1 - P)
> >
> > for each P and rms in both files.
> >
> > You will find that the indice can be reconstructed, (differing only by
> > a scaling constant,) from its statistics-a validation of the model
> > used.
> >
> >         John
> >
> > BTW, there are a lot of tricks that can be used from the
> > http://www.johncon.com/ntropix/utilities.html page. The tsgainwindow
> > does about the same thing, but a geometic progression has to be done
> > on the output.
> >
> > John Conover writes:
> > > BTW, another interesting thing to do is:
> > >
> > >     tsshannonwindow -w 100 -a -b -c -d -e -f -g -h data_file
> > >
> > > which calculates the Shannon entropy, using different methods. Vary
> > > the -w argument.
> > >
> > > The first thing one notices is that the Shannon entropy is fractal,
> > > (unless the -w argument is set to many thousands-consistent with the
> > > tsshannoneffective program,) and all the methods give values that
> > > are very close to the theoretical value:
> > >
> > >     P = ((avg / rms) + 1) / 2
> > >
> > > even though some of the methods don't measure the avg or rms at all,
> > > (for example, some just count the number of ups, and downs, and
> > > compute the entropy from ups / (downs + ups) in the time interval
> > > defined by the -w argument.) Others assume that the absolute value of
> > > the movements are the same as the rms.
> > >
> > > Its kind of an interesting exercise.
> > >
> > >         John
> > >
> > > BTW, the manual page for the tsshannonwindow program is at:
> > > http://www.johncon.com/ndustrix/archive/utilities/tsshannonwindow.txt, and the
> > > output is a standard Unix tab delimited file. So, use "cut -f1,2" and
> > > "cut -f1,3" and so on to get different files of the different methods
> > > of calculating the Shannon probability. Its kind of interesting to
> > > make a plot of them overlayed. The gist of it is that stocks increase
> > > in value, (or decrease,) more based on the ratio of the number of up
> > > movements to down movements in an interval than by the magnitude of
> > > the movements in an interval. Kind of counter intuitive, (and a good
> > > empirical statement of the fractal nature of such things.)
> > >
> > > John Conover writes:
> > > > Just as kind of an FYI, Blake LeBaron,
> > > > (http://www.ssc.wisc.edu/~blebaron/,) one of the NLDS, (chaos,)
> > > > theorist pointed out in 1991, that for some reason, market crashes are
> > > > always preceded by an increase in the root mean square of the daily
> > > > marginal returns of the indices. (Which is vary characteristic of
> > > > bifurcations in NLDSs.)
> > > >
> > > > If you use the tsrmswindow program, (from
> > > > http://www.johncon.com/ntropix/utilities.html,) on the historical
> > > > database of the DJIA, S&P500, and NASDAQ, it seems to be true. For
> > > > example:
> > > >
> > > >    tsfraction data_file | tsrmswindow -w 100
> > > >
> > > > where data_file is the time series for the NASDAQ, will make a plot
> > > > file that shows that for several years, the rms values have been
> > > > running about 3X their average value-averaged since 1971. A like
> > > > scenario happened in the late 20's to the DJIA and S&P. Likewise for
> > > > the other crashes and crash'ets of the 20'th century.
> > > >
> > > >         John
> > > >
> > > > BTW, as nearly as I can tell, the US equity market has degenerated
> > > > enough such that 5-10% of the US's net wealth has went up in smoke.
> > > > About 3-5 trillion bucks have been lost, (depending on who is doing
> > > > the counting,) and the US net wealth is estimated at about 50 trillion
> > > > bucks-about a forth to half of the world's net wealth, (Re: the US
> > > > FED-I have no idea how they measure that; what is the value of the
> > > > nuclear weapons arsenal? How is it depreciated?) Its a fairly sizable
> > > > chunk of the world's net wealth.
> > > >

--

John Conover, john@email.johncon.com, http://www.johncon.com/


Copyright © 2001 John Conover, john@email.johncon.com. All Rights Reserved.
Last modified: Fri Jan 5 22:28:03 PST 2001 $Id: 010103015826.6705.html,v 1.0 2001/11/17 23:05:50 conover Exp $
Valid HTML 4.0!