Subject: Re: What is the difference between chaotic and random?
From: "Robert. Fung"
Date: Wed, 13 Nov 1996 15:27:37 -0500
Robert Dodier wrote:
>
> Hello all,
> It's a beautiful blue day here in Boulder, hope it's the same whereever
> you are.
>
> Troy Shinbrot wrote:
> >
> > In article <19961108025700.VAA15372@ladder01.news.aol.com>,
> > jksnyder@aol.com wrote: > >
> > > Please excuse the elementary nature of the question, I am just learning
> > > about chaotic systems. Is there a difference between chaotic systems and
> > > random systems? If so, could the difference be measured/quantified by
> > > plotting the series on normal probability paper?
> > >
> > > I generated a chaotic series of 1,000 numbers, between zero and one,
> > > using the logistic equation and similarly generated a series of 1,000
> > > numbers, between zero and one, with a random number generator. After
> > > ordering both series, they both plotted as straight lines on normal
> > > probability paper. Therefore, by this test, chaotic and random number
> > > series appear to have a normal distribution.
>
> This seems very strange; the invariant measure for ax(1-x) is far from
> normal. For a=4, the invariant measure is continuously differentiable
> and thus there is a density function, which looks like a U. For a < 4,
> the invariant measure has a lot of spikes (I don't recall if the number
> of spikes is finite, countable, or uncountable), so there is no density.
> Also, when you say a random number between 0 and 1, I believe you must
> mean uniformly distributed; again, this is anything but normal. If you
> make a histogram of the 1000 numbers, what shape do you get?
>
> > > Are there other measures, such as analysis of variance, that could
> > > distinguish between random and chaotic series?
>
> I'll try to argue that this is, for practical purposes, not a meaningful
> question. First, let's review what Mr. Shinbrot wrote. By the way, this
> is
> indeed a great question, which touches on a fundamental issue.
>
> > A great question. First a practical answer for this particular problem.
> > If X(n) denotes the n-th value of your time series, if you plot X(n) vs.
> > X(n-1), you will get a mess for the random data, because the n-th value of
> > the time series does not depend on the n-1'st value. For the logistic
> > data, you will get a parabola (obviously).
> >
> > Thus because of the deterministic nature of chaos, one value depends on
> > its history, while random data does not. Vis a vis statistical tests, if
> > you randomize the ORDER of the logistic data, you will have two data sets,
> > one logistic and a second randomized-order logistic, both of which are
> > guaranteed to have EXACTLY the same mean, variance, skew, kurtosis, or
> > anything else you would care to measure. It is only the determinism of
> > the data sets that differ.
>
> I don't think dependence on the past is a suitable way to distinguish
> random from chaotic processes. I'm sure we'll all agree that a Markov
> process is a random process, yet such a process may have a very strong
> dependence on past states.
>
> > The second answer is more pedagogical: the logistic data have dimension at
> > most 1. That is, they all lie precisely on the parabola I mentioned, and
> > one variable is all that is needed to define the state and thus determine
> > the future state of the system. The random data are (ideally) infinite
> > dimensional: an ideal random number generator would require an infinite
> > number of variables to define the future state. Practically we all know
> > that this isn't quite true, but that is the strict answer.
>
> There is nothing in the textbook definition of a random variable that
> requires that it be generated by an infinite-dimensional problem. In
> order
> for all the definitions about expected value, distribution function,
> density, etc etc to work, all that is required is that the generating
> process have a unique invariant measure, so that time averages over
> process values equal weighted averages taken over the state space (with
> the invariant measure doing the weighting). That is, a random variable
> need not be generated by an infinite-dimensional process; the process
> need
> only be ergodic -- this is a much weaker condition.
>
> Incidentally, for ergodic processes the frequentist and
> measure-theoretic
> definitions of probability coincide. I don't think the followers of
> these
> two schools really differ on any practical point.
>
> So I've pointed out that random processes can be low-dimensional, but
> I could make the argument a little more convincing by coming up with
> some examples of deterministic system which has an everyday distribution
> as its invariant measure. So far I can't think of a low-dimensional
> system
> which has an invariant measure which is approximately normal, say.
> Can anyone name such a system?
>
For a small finite dimensional state space like a single die throw, one
can devise an infinite number of deterministic machines to throw the die
with differing levels of randomness while maintaing a flat distribution across
each of the 6 states, over a projected infinite number of trials.
If one machine "A" throws 1,2,3,4,5,6,1,2,3,4,5,6.... and
the another "B" simularly throws 1,2,3,4,4,3,6,5,1,2,5,6,1,2,... what is the measure
that says "B" is more random than "A" ? Both are deterministic and
have flat distributions of their states.
Should a random process be proveably non-deterministic ?
Subject: Re: Occam's razor & WDB2T [was Decidability question]
From: maj@waikato.ac.nz (Murray Jorgensen)
Date: Thu, 14 Nov 96 16:06:08 GMT
I regret that I do not have the time to respond to this thread in detail.
I have looked at Geoff Webb's article in
http://www.cs.washington.edu/research/jair/table-of-contents-vol4.html
and it seems to conflict with all my intuition built up as a practising
statistician.
The subject of 'machine learning' is very closely connected with the
fitting of statistical models to empirical data. What the ML people have
contributed is a range of new algorithms and models but the fundamental
questions remain unchanged. It is widely accepted in the statistical
community that 'overfitting' of a data set [using a needlessly complex
model] results in a fitted model closely tuned to that particular data
set that has poor predictive power. This is not to say that there is not
additional complexity to be discovered, just that the data set under
consideration does not contain enough information about possible
elaborations to the model to make it safe to fit them.
I recommend the book
Model Selection by H. Linhart and W. Zucchini
Wiley 1986 ISBN 0-471-83722-9
Murray Jorgensen
In article <32837820.7ACB@postoffice.worldnet.att.net>,
kenneth paul collins wrote:
>kenneth paul collins wrote:
>
>> From the view of WDB2T, Occam's Razor can be sharpened a bit. In
>> terms of WDB2T, the more-complex alternative simply does not fit
>> observations well, and it can be rejected solely on that basis.
>> And when one looks, one sees that this is the the same point that
>> I've been working to make with respect to the relative utility of
>> conventional Logic and this "new" WDB2T-optimization "Logic" I am
>> proposing.
>
>[For those who didn't read the "Decidability" thread, "WDB2T" is an
>acronym for "What's Described By the 2nd law of Thermodynamics".
>WDB2T refers to the Physical Reality that is described by 2nd Thermo,
>not 2nd Thermo itself.]
>
>Yesterday, I came across a short, unsigned, report in the Nov 96
>issue of _Discover_ magazine, p34, "Is Occam's Razor Rusty?". The
>article reports on work done by Geoffrey Webb at Deakin University in
>Geelong, Austrailia. In a series of experiments, Webb found that,
>(quoting from the _Discover_ article) "for 12 of 13 problems analyzed
>by the computer, the more complex decision-making process gave more
>accurate results".
>
>The _Discover_ article quotes Webb: "'People are potentially missing
>out on useful patterns because they're just looking for the simple
>ones,' says Webb. 'Occam's razor influences and limits what science
>can do with information."
>
>The article ends without clarifying the point, but my interpretation
>is that it's (Webb is) saying that Occam's razor =erroneously=
>"influences and limits what science can do with information", and
>since this contradicts the position that I've recently taken here in
>sci.logic with respect to Occam's razor, I wish to explore this
>matter further.
>
>This msg is an introduction to the new thread. I'll post further
>discussion later today. ken collins
>_____________________________________________________
>People hate because they fear, and they fear because
>they do not understand, and they do not understand
>because hating is less work than understanding.
Subject: Re: Pronunciation of LaTeX
From: buhr@stat.wisc.edu (Kevin Buhr)
Date: 13 Nov 1996 15:45:38 -0600
-----BEGIN PGP SIGNED MESSAGE-----
Hideo Hirose writes:
|
| In Japan, many researchers pronounce LaTeX as "latef." Is it
| correct? How do you pronounce TeX and LaTeX actually, especially in
| the united states?
Donald E. Knuth spends an entire chapter of the TeXbook on the
pronounciation of "TeX". Okay, the "chapter" is only one page long,
and he talks about some other stuff, too. However, it is abundantly
clear that the author of "TeX" wants you to pronounce it "teck" (i.e.,
rhymes with "blech"), not "tecks". Evidently, "when you say it
correctly to your computer, the terminal may become slightly moist".
If I remember correctly, Leslie Lamport makes it clear that there is
*no* official pronounciation of "LaTeX". He recommends one of:
(1) lah-TECK
(2) LAH-teck
(3) LAY-teks
However, on any CTAN site, the "CTAN/latex/intro.tex" document will
tell you that LaTeX should be pronounced as one of:
(1) Lah-tech
(2) Lay-tech
(in each case, rhyming with blech).
So, the correct solution is to avoid pronouncing LaTeX whenever
possible. And if you must pronounce it, try to mumble.
Kevin
-----BEGIN PGP SIGNATURE-----
Version: 2.6.3
Charset: noconv
Comment: Processed by Mailcrypt 3.4, an Emacs/PGP interface
iQBVAwUBMopBW4mVIQW1OgXhAQHJ/wIAhOCzstNW3GLyvA8OG3vmh4IV198NgwY3
STX/mxnGmhLdIrfK1FEWmT6sf6ae56exTUmODdKRvbKZ+Ou7qFxYaA==
=ZN8k
-----END PGP SIGNATURE-----
Subject: Re: What is the expectation of the distance between 2 points in a unit square?
From: rbcrosie@apgea.army.mil (Ronald B. Crosier)
Date: Thu, 14 Nov 96 13:44:13 GMT
In article <56csun$85@corn.cso.niu.edu>, Rick Johns wrote:
>This is a problem passed to me by another. Although I can figure out
>what the mean distance between two points in a unit square is by
>empirical sampling, I would like to know how to express the expectation
>mathematically. Can anyone help with this? I am also interested in
>having a general solution for higher dimensions as well. Thanks in
>advance.
>
I sent Rick Johns an old post by Steve Finch (from sci.math) that
contains the answer he (Rick) wants. Part of Steve's post follows.
--
Ghosh also provides formulas for the moments of the ... distribution. In the
special case of a square with unit sides (a=b=1), the mean distance between
the two random points is
{sqrt(2) + 2 + 5 log(1+sqrt(2))}/15,
which is approximately 0.52141, ... .
Formulas for mean distances associated with other convex planar sets are
given in Santalo[2]. As far as I know, no analogous exact formulas exist
for convex sets in higher dimensions - simulation may indeed be the only
approach to determine distributions/moments for, e.g., the cube or the sphere.
References
[1] B. Ghosh, Random distances within a rectangle and between two rectangles,
Bull. Calcutta Math. Soc., 43 (1951) 17-24.
[2] L. Santalo, Integral Geometry and Geometric Probability, Addision-Wesley,
1976, p.49.
At least two more recent books on stochastic geometry have recently appeared
but I cannot recall further information about them. ... Steve Finch
--
Ronald Crosier E-mail:
Disclaimer: My opinions are just that---mine, and opinions.
If you have a good idea, be patient---it will go away.