Back


Newsgroup sci.stat.consult 21208

Directory

Subject: box-cox transformation -- From: engp6373@leonis.nus.sg (Than Su Ee)
Subject: pearson distribution -- From: engp6373@leonis.nus.sg (Than Su Ee)
Subject: CONJOINT-analysis in SAS -- From: Roy Stewart
Subject: Help Gelatin -- From: craffae1@extro.ucc.su.oz.au
Subject: Re: Cronbach alpha -- From: Rick DeShon
Subject: URGENT gelatine -- From: craffae1@extro.ucc.su.oz.au
Subject: help: robust estimation -- From: Roel.Leus@student.kuleuven.ac.be (Roel Leus)
Subject: The "crud" factor! [Meehl] -- From: rwhite@superior.carleton.ca (Robert White)
Subject: Re: Cronbach alpha -- From: Dennis Roberts
Subject: Seeking statistician for software development in Cambridge MA -- From: Tom Lane
Subject: Testing one two three -- From: "Aufgang"
Subject: Can you please help? -- From: Daniel Aufgang
Subject: Re: Adequacy of sample size -- From: Dennis Roberts
Subject: Help with Probability on Temperture variations -- From: Jon Sauvageau
Subject: Re: box-cox transformation -- From: "Richard D. McCleary"
Subject: Re: E[ X | X => X^*] = ? -- From: Hans-Peter Piepho
Subject: Sample Size for Estimating Regression Equation -- From: Zubin Dowlaty
Subject: Basic Asymtotics -- From: Tatsuo Ochiai
Subject: Basic Asymtotics -- From: Tatsuo Ochiai
Subject: LMedS regression -- From: waynet@pendragon.cna.tek.com (Wayne Turner)
Subject: Re: bootstrap methods -- From: gregorich@psg.ucsf.edu (Steve Gregorich)
Subject: Re: Power? -- From: wpilib+@pitt.edu (Richard F Ulrich)
Subject: Re: Tricky sampling question -- From: T.Moore@massey.ac.nz (Terry Moore)
Subject: Capture and re-capture data -- From: Chung Sai Ho
Subject: Re: Tricky sampling question -- From: mdupin@kodak.com (Mike)
Subject: I found out how to use fft data in MS excel -- From: peter homel
Subject: Statistics of outcomes of competitions -- From: Eric Bartels
Subject: Re: prediction contest -- From: Gerrit.Jacobsen@hamburg.netsurf.de (Gerrit Jacobsen)

Articles

Subject: box-cox transformation
From: engp6373@leonis.nus.sg (Than Su Ee)
Date: 26 Nov 1996 10:37:05 GMT
I need a ftn77 porgram or subroutine that is able to perform box-cox 
transfomartion.  IMSL's box-cox subroutine require the user to input the 
power index (lambda) calculated from maximizing log likelihood function.  
I am wondering whether there is subroutine that incorporates both the 
lambda-seeking and transforming features. Thanks
Return to Top
Subject: pearson distribution
From: engp6373@leonis.nus.sg (Than Su Ee)
Date: 26 Nov 1996 11:50:29 GMT
I need a fortran77 subroutine to calculate the percentage point for 
pearson distribution.  If you have any information on where to get this, 
do let me know.  Really appreciate.
*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*
 Than Su Ee                                 Tel:(65) 7722208(Off)	
 Dept. of Industrial & Systems Engineering  Fax:(65) 7771434
 National University of Singapore      	    Email1:engp6373@leonis.nus.sg
 10 Kent Ridge Crescent                     Email2:suee@post1.com 
 Singapore 119260                            
*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*
Return to Top
Subject: CONJOINT-analysis in SAS
From: Roy Stewart
Date: Tue, 26 Nov 1996 13:26:38 -0800
Does any one have any suggestions
HOW to do a CONJOINT-analysis in SAS
THANKS for any response
Roy Stewart
e-mail: R.Stewart@MED.RUG.NL
Return to Top
Subject: Help Gelatin
From: craffae1@extro.ucc.su.oz.au
Date: Wed, 27 Nov 1996 00:14:47 +1000
URGENT URGENT URGENT
I am doing some URGENT (ie due on 29/11) research on packet UNflavoured
(this is crucial there is enough info on flavoured variety) Gelatine. 
It is very hard to find statistics on its use (ie what are the buying
trends) especially in Europe or Asia.
If anyone can help me in the time deadline I will be ever so grateful.
If anyone has any information or leads where I could find this
information please please email me at craffae1@extro.ucc.su.oz.au ASAP. 
The character before the @ is a #one, 1, not a letter l.
Please email me instead of posting a reply as I will not be checking
this newsgroup.
Thanks!
Catherine Raffaele
Return to Top
Subject: Re: Cronbach alpha
From: Rick DeShon
Date: Tue, 26 Nov 1996 08:01:54 -0500
Dennis Roberts wrote:
>Cronbachs alpha is not really a type of correlation ... but rather an index
>of internal consistency reliability. One reference would be: Hoi Suen
>(1990), Principle of test theories, Lawrence Erlbaum, Hillsdale: NJ ... see
>pages 35, 57, and 61.
Although alpha is used as an index of internal consistency it is a type of
correlation.  Specifically, it is an intra-class correlation.  Shrout and
Fleiss (1979) Psychological Bulletin and McGraw and Wong (1996)
Psychological Methods highlight this interpretation.
Rick
Richard DeShon
Dept. of Psychology
Michigan State University
East Lansing, MI  48824-1117
E-mail: deshon@pilot.msu.edu
Voice:  (517) 353-4624
Fax:    (517) 353-4873
Return to Top
Subject: URGENT gelatine
From: craffae1@extro.ucc.su.oz.au
Date: Tue, 26 Nov 1996 07:01:06 -0600
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
URGENT URGENT URGENT
I am doing some URGENT (ie due on 29/11) research on packet UNflavoured (this is 
crucial there is enough info on flavoured variety) Gelatine.  It is very hard to 
find statistics on its use (ie what are the buying trends) especially in Europe 
or Asia.
If anyone can help me in the time deadline I will be ever so grateful.
If anyone has any information or leads where I could find this information 
please please email me at craffae1@extro.ucc.su.oz.au ASAP.  The character 
before the @ is a #one, 1, not a letter l.
Please email me instead of posting a reply as I will not be checking this 
newsgroup.
Thanks!
Catherine Raffaele
PS I don't know how to use newsgroups so please excuse the newbie bumbling!
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
-------------------==== Posted via Deja News ====-----------------------
      http://www.dejanews.com/     Search, Read, Post to Usenet
Return to Top
Subject: help: robust estimation
From: Roel.Leus@student.kuleuven.ac.be (Roel Leus)
Date: 26 Nov 1996 13:40:48 GMT
I am a fourth year student of Trade Engineer at the KU Leuven.
In the framework of a paper I am to write on detecting outliers
in multivariate samples, I am at the moment looking into robust
estimation of multivariate location and shape, with which I could
calculate the Mahalanobis-distances of each point to the center.
I already found : MVE, MCD, M- and S-estimators.
Are there other suggestions, and what (basic) books and articles
should I consult ?
Return to Top
Subject: The "crud" factor! [Meehl]
From: rwhite@superior.carleton.ca (Robert White)
Date: 26 Nov 96 14:17:42 GMT
Pervin, L.A.(1990)_Handbook of Personality Theory & Research_
                  The Guilford Press New York
          "As Meehl [in press] notes, our fascination with
           correlations and tests of statistical significance
           often leads us to assume progress in understanding
           a phenomenon when none is present; this is due in part
           to what he calls the "crud" factor. The history of
           fads in personality research, noted in the intro-
           ductory chapter, is undoubtedly relevant here." p724
Date: 25-Nov-96
Database: PsycINFO <1984 to November 1996>
Set   Search                                                       Results
---------------------------------------------------------------------------
005   crud factor.tw.                                                    2
<1>
Accession Number
  Journal Article:  79-18428.
Authors
  Standing, Lionel;  Sproule, Robert;  Khouzam, Nelly.
Institution
  Bishop's U, Lennoxville, PQ, Canada.
Title
  Empirical statistics: IV. Illustrating Meehl's sixth law of soft
  psychology: Everything correlates with everything.
Source
  Psychological Reports. Vol 69(1) 123-126, Aug 1991.
Abstract
  Computed a 135 * 135 matrix of correlations, using
  educational/biographical data on 2,058 Canadian grade-school children, to
  find the background level of statistically significant correlations, or P.
  Meehl's (1990) "crud factor." All of the 135 variables, except student
  identification number, displayed more statistically significant
  correlations with the other variables than could be predicted from chance.
  Results reinforce Meehl's criticisms of overreliance on significance
  levels for correlational data. (PsycINFO Database Copyright 1992 American
  Psychological Assn, all rights reserved).
<2>
Accession Number
  Chapter:  91-011021-001.
Authors
  Meehl, Paul E.
Institution
  U Minnesota, Professor of Psychology, MN, US.
Chapter Title
  Why summaries of research on psychological theories are often
  uninterpretable. [References].
Book Citation
  Improving inquiry in social science:  A volume in honor of Lee J.
  Cronbach. (Richard E. Snow, David E. Wiley, Eds.), pp. 13-59.
Source
  Lawrence Erlbaum Associates, Inc, Hillsdale, NJ, US; xiv, 423 pp. 1991.
Content Representation
  (from the book)  begins ...with a general rejection of the hypothesis
  testing tradition that has dominated so much of psychology and social
  science # argues that most of our research literatures are rendered
  uninterpretable because of a series of obfuscating factors the effects of
  which are often sizable, variable, operating in opposite directions, and
  left unmeasured # urges that we come to full appreciation of how little we
  really know, and how little can ever be learned through null hypothesis
  testing. (from the chapter)  (discusses) surveys of research evidence
  sharing three properties ...(a) theories in so called " soft areas,"
  (namely, clinical, counseling, personality theory, and social psychology),
  (b) data correlational, and (c) positive findings consisting of refutation
  of the null hypothesis /// ten obfuscating factors that make (null
  hypothesis) refutation in the soft areas largely uninterpretable # loose
  derivation chain # problematic auxiliary theories # problematic ceteris
  paribus clause # experimenter error # inadequate statistical power # crud
  factor # pilot studies # selective bias in submitting reports # selective
  editorial bias # detached validation claim for psychometric instruments.
---------------------------------------------------------------------------
--
   ----------------------------------------- Carleton University ----------
               Robert G. White               Dept. of Psychology
                                             Ottawa, Ontario. CANADA
   INTERNET ADDRESS ----- rwhite@ccs.carleton.ca ------------------- E-MAIL
   ------------------------------------------------------------------------
Return to Top
Subject: Re: Cronbach alpha
From: Dennis Roberts
Date: Mon, 25 Nov 1996 23:38:36 -0500
Not to beat this to death but ... alpha may be one application of the intra
class correlation but ... correlations of this type are not restricted to
estimates of reliabilty ... so, in that sense, I think it is better not to
think of it AS a type of correlation ...
When we talk about "types" of rs ... we tend to think of Pearson r ... or
Spearman rho ... or a tetrachoric r ... etc. .... and Cronbach's alpha seens
not to be in the same spirit of "type" as these.
Another way to describe this is ... when we talk in a stat course in a
chapter on rs ... we would never bring up Cronbachs alpha ...
At 08:01 AM 11/26/96 -0500, you wrote:
>Dennis Roberts wrote:
>
>>Cronbachs alpha is not really a type of correlation ... but rather an index
>>of internal consistency reliability. One reference would be: Hoi Suen
>>(1990), Principle of test theories, Lawrence Erlbaum, Hillsdale: NJ ... see
>>pages 35, 57, and 61.
>
>Although alpha is used as an index of internal consistency it is a type of
>correlation.  Specifically, it is an intra-class correlation.  Shrout and
>Fleiss (1979) Psychological Bulletin and McGraw and Wong (1996)
>Psychological Methods highlight this interpretation.
>
>Rick
>
>
>Richard DeShon
>Dept. of Psychology
>Michigan State University
>East Lansing, MI  48824-1117
>E-mail: deshon@pilot.msu.edu
>Voice:  (517) 353-4624
>Fax:    (517) 353-4873
>
>
===========================
 Dennis Roberts, Professor EdPsy             !!! GO NITTANY LIONS !!!
 208 Cedar, Penn State, University Park, PA 16802 AC 814-863-2401
 WEB (personal) http://www2.ed.psu.edu/espse/staff/droberts/drober~1.htm
Return to Top
Subject: Seeking statistician for software development in Cambridge MA
From: Tom Lane
Date: Tue, 26 Nov 1996 11:15:40 -0500
Domain Solutions Corporation is seeking a statistician to help us
develop software for analyzing data in a manufacturing environment.
This statistician will
    - Define the statistical content of software releases.
    - Assist the quality assurance group in designing, carrying
      out, and validating tests.
    - Write software using C++ for Unix and Windows.
In addition, this person may occasionally be called upon to write
technical documentation, consult with software engineers, perform
competitive analysis, support the sales organization, and interact
with customers.
Qualifications:
  Required:
    - Masters in statistics.
    - Knowledge of applied statistics areas such as regression,
      quality control, design of experiments.
    - Experience in software development.
  Preferred:
    - Ph.D. in statistics or a related field.
    - Experience in software testing.
    - Experience applying statistics in a manufacturing environment.
Domain Solutions Corporation, formerly a division of BBN, develops
and markets software to optimize manufacturing processes and to
manage clinical trials data.  Our products include RS/1,
RS/Explore, RS/Discover, Cornerstone, Probe, Patterns, and Starfire.
You can apply for this position by sending a cover letter and resume
via email to tlane@domaincorp.com, by fax to 617-873-8199, or by
mail to Tom Lane, Chief Statistician, Domain Solutions Corporation,
150 CambridgePark Dr., Cambridge, MA 02140.  Please send email in
the form of plain text, postscript, or Microsoft Word.
Return to Top
Subject: Testing one two three
From: "Aufgang"
Date: 27 Nov 1996 15:30:47 GMT
Ugggg!!!!!
Return to Top
Subject: Can you please help?
From: Daniel Aufgang
Date: Tue, 26 Nov 1996 18:08:09 +0200
Please send answers to: aufgang@netmedia.net.il
Hello.  First off, I am sorry about any cross posts and if this letter
is not relevant for this list. I am finishing up my thesis on "The
Subjective, Objective and Behavioral Environment in Urban Planning" and
am now stumped on the statistics part.  In short, I know I have to use
multi dimensional scaling, but I do not know about all the sub-types
and their nuances.  I feel, like factor analysis, that if you do not
know the methods inside and out, you are run into trouble when "those
who do" read over your =FDwork.  I am using Statistica 5.0 and Excel for
the analysis.=20
Here is my problem.  I conducted an open-ended survey which I am using
to conduct a close ended survey.  I asked questions like "What words
would you use to describe the =FDlayout of the mall, the look of the
mall.." etc.  In each question category, I compiled =FDabout 20-50 words.=
=20
Here are the responses to one of the questions; What is the main reason
you came here today.  (translated from Hebrew for obvious reasons)
-to buy certain products
-movies
-window shopping
-general shopping
-eat
-drink coffee
-get out of the house
-center for entertainment
-walk around
-trip
-get services
-to do everything at once
-bank
-to enjoy the atmosphere
-meet friends
-pass the time
-to see things
-entertainment
-meet girls
-for no reason
-play
-rest
-for the children
-see what's new
-get bored
I put them into the following groups:
        -meet friends
-> I am using the Kenyon as a place to meet my friends.
        -general shopping
        -to buy certain products
        -get services
        -to do everything at once
        -bank
->I came here to do my shopping
        -entertainment
        -pass the time
        -to see things
        -window shopping
        -get out of the house
        -center for entertainment
        -see what's new
        -trip
        -for the children
        -to enjoy the atmosphere
        -rest
        -play
        -meet girls
->I came here to have an enjoyable time.
        -drink coffee
        -I came here to eat
->I came here for the resturaunts and cafes.
        -movies
->I came here to to go a movie
I then asked 20 people: "These are reasons why people claimed they were
in the Kenyon HaArem the day of our survey. Please group the reasons
into logical groups. Try to create between 3 to  groups." The groups
differed from what I chose. For example, one student put "To get
services" & "bank" in their own category of SERVICES and "to buy certain
products" "general shopping" "drink coffee" "eat" and "to do everything
at once" under the category "To perform specific functions"
I then made a matrix which shows how often  each word appears with each
other. (triangle matrix)
Now what? Is my question worded ok?  What is the best way of creating
the matrix? I did it in Excel and it took my a very long time.  I will
soon have MANY MANY more questionnaires to enter. What is the most
valid way of grouping this data?  In the end, I want groups like I
created.  For the other questions, these will be turned into Personal
Constructs.
Return to Top
Subject: Re: Adequacy of sample size
From: Dennis Roberts
Date: Mon, 25 Nov 1996 23:33:51 -0500
The fact is ... this IS complicated. In your case, to make it as simple as
possible .. you are looking for a correlation between some predictor and
some criterion. Well ... how BIG should that be .... for you to think the
relationship is strong enough for you to conclude that there is an IMPORTANT
relationship between X and Y? This is the first and most fundamental
question that has to be answered ... and the answer is NOT ... big enough to
be significant. Then ... how relatively critical would it be to make errors
of the type I or type II variety ... ie, is it relatively more serious (or
equal) to conclude that there is a relationship when perhaps there is not
... or to conclude there is NO relationship when in fact there is?
Only after you answer a series of questions like the above ... can you then
start to attempt to answer the question: how large of a sample do I need to
find a relationship as big as I have said is important ... given that I am
willing to tolerate some percent of making each of the errors that can be
made ...
As I said ... it ain't simple.
At 10:14 PM 11/25/96 GMT, you wrote:
>I've asked in another post how one tells how large a sample size to
>use to get significant results.  The responses I've gotten include:
>(1) ask a statistician, and (2) it's too complicated to tell without
>knowing what you're trying to prove.  So here's my research question
>and how I intend to go about answering it.  If you have the time, I
>welcome any suggestions--especially with regard to how many patients
>I'll need to include in the study to get significant results.  Thanks.
>
>At my hospital (where I'm a resident) we admit patients through the
>emergency room to the family practice unit.  Some of those patients
>come in with respiratory diagnoses (such as pneumonia and asthma).
>Occasionally, a patient's condition will deteriorate, and he'll have
>to be transferred to the medical intensive care unit (MICU) for more
>intensive care (such as being placed on a ventilator).  This is true
>for patients with all conditions, but my study only deals with
>respiratory conditions.
>
>It would be in the interests of better patient care if we could
>identify in the emergency room (before the patient is ever admitted to
>the hospital) which respiratory patients are likely to be transferred
>to the MICU.  That way, we could admit them directly to the MICU or,
>more commonly, place them in an intermediate-care unit (what we call a
>"step-down" unit).
>
>So the research question is:  which respiratory patients admitted to
>the family practice unit are likely to be get worse rather than
>better, to the point of needing transfer to the MICU?
>
>We have identified about 20 variables which we would like to test.
>These include age, gender, blood oxygenation (how much oxygen is
>actually getting into the body through the diseased lungs),
>co-existing medical conditions, previous admissions to the hospital
>for respiratory problems, previous admissions to the MICU for
>respiratory problems, and many other factors.
>
>The data will be gathered as follows:  When the patients are admitted
>to the family practice unit, a researcher will gather the values of
>each of the twenty variables.  These values will be obtained from the
>chart and, if necessary, from the patient.  The researcher will not be
>anyone involved in the patient's care.
>
>After the patient has been discharged from the hospital, the
>researcher will pull the patient's chart and find out whether the
>patient went to the MICU.  This will be the outcome variable.
>
>After data on a sufficient number of patients is gathered, the SAS
>program will be used to figure out which if any of the variables are
>predictors of decline in respiratory patients.  Those variables which
>are predictors will then be incorporated into a formula whose result
>will be an "index of decline in respiratory patients admitted to the
>family practice unit through the emergency room".
>
>In the second phase of the study, data will be gathered in a similar
>manner, with the exception that only those variables incorporated into
>the index will be needed.  When data on a sufficient number of
>patients is gathered, the SAS program will be used to look for a
>correlation between the index and the outcome.
>
>Thanks for evaluating my study design.  Your suggestions are welcome
>and appreciated.
>
>Also, I'd be most interested in any explanations you care to offer
>regarding, in a general way, how one goes about determining adequacy
>of sample size.
>
>Matt Beckwith, M.D.
>Jacksonville, Florida
>
>
===========================
 Dennis Roberts, Professor EdPsy             !!! GO NITTANY LIONS !!!
 208 Cedar, Penn State, University Park, PA 16802 AC 814-863-2401
 WEB (personal) http://www2.ed.psu.edu/espse/staff/droberts/drober~1.htm
Return to Top
Subject: Help with Probability on Temperture variations
From: Jon Sauvageau
Date: Tue, 26 Nov 1996 11:10:22 -0800
In article <329A5D82.2741@gsfc.nasa.gov> you write:
>R wrote:
>> 
>> In article <3291F61B.E13@gsfc.nasa.gov>, Jon Sauvageau  says:
>> >
>> >Currently I am doing a small school project that deals with temperature variences inside a building vs what the outside temp is doing in relation to the cost of energy (electricity).  Has anyone done something similar and would share your data/process.  I need help in setting up tests such as Chi Square or some other hypothesis test.  If you can help in anyway, point me to a BB or maybe suggest a book/artical.  Thanks in advance.
>> >
>> >Jon Sauvageau
>> 
>> If you can find the book "Facts From Figures" by M.J. Moroney, which is a paper back and quite cheap, I think this will help enormously.
>> R.
>Can you help me with this,  I have looked in several libraries and book
>stores to no avail.  Do you have a source for this book?
>
>Please let me know ASAP.
>
>Jon Sauvageau
Last year someone in sci.stat.consult posted a similar request.
If you post your request there, you might get some real help.
Tony Rizzo
If you know of this book or know where I can locate more info please let
me know.   Jon.M.Sauvageau.1@gsfc.nasa.gov
Return to Top
Subject: Re: box-cox transformation
From: "Richard D. McCleary"
Date: Tue, 26 Nov 1996 08:56:01 -0800
The Box-Cox fnctn has a very simple likelihood (after a simple rescaling).
See Kmenta's *Elements of Econometrics, 2nd Ed.* (Macmillan, 1986), pp.
517-521, for this material.  You can program a ML routine in a few steps.
Return to Top
Subject: Re: E[ X | X => X^*] = ?
From: Hans-Peter Piepho
Date: Tue, 26 Nov 1996 15:30:52 +0100
>Tatsuo  asked:
>
>>Suppose X ~ N(mu, sigma=AC2). Then, what is the formula for
>>     E[X|X=3D>X=AC*] : Expected value of X given X is greater than or
>>equal      to some fixed number X=AC*
>
>you can find the expected value by evaluating the integral of x*p(x) from
> -mu/sigma=AC2  to oo  and multiplying the result with sigma=AC2
>p(x) is the pdf of N(0,1) and is equal to  exp(-x*x/2)*(2*pi)=AC-0.5
>
>BTW the indefinite integral of x*p(x) in this case evaluates to -p(x)
>
>The book that I found most thorough in explaining Expected values and
>introductory statistics in general was :
>"Statistical Theory and Methodology in Science and Engineering"
>by K.A.Brownlee, 1960
>In my opinion this book beats any modern intro statistics book hands-down.
>Page 38 is on expected values.
>
>I had asked a similar question in this newsgroup and it took me a good
>week to justify the answer I got but I finally did. Thanks again to Aaron
>Brown who gave me the initial help.
>
>Dimitri
>
>
This is the expected value of a truncated normal with truncation from below=
=20
at X. The truncated normqal is discussed in: Johnson and Kotz 1970=20
Continuous univariate distributions I. p.81
Let f(z) denote the p.d.f of the standard normal and F(z) its c.d.f.
E(X|X>X*) =3D mu + sigmaµ2*f(Z*)/(1-F(Z*))
where Z* =3D (X*-mu)/sigma
_______________________________________________________________________
Hans-Peter Piepho
Institut f. Nutzpflanzenkunde  WWW:   http://www.wiz.uni-kassel.de/fts/
Universitaet Kassel            Mail:  piepho@wiz.uni-kassel.de
Steinstrasse 19                Fax:   +49 5542 98 1230
37213 Witzenhausen, Germany    Phone: +49 5542 98 1248
       =20
                                    =20
Return to Top
Subject: Sample Size for Estimating Regression Equation
From: Zubin Dowlaty
Date: Tue, 26 Nov 1996 15:20:52 -0500
How does one go about calculating the desired sample size when one is
estimating a linear regression from the data?  
	Should I estimate(guess) the variability of the coefficents and then
plug these variance estimators into the standard sample size formula for
determining the desired sample for a population mean(with a predefined
tolerable width setting)??
Return to Top
Subject: Basic Asymtotics
From: Tatsuo Ochiai
Date: Tue, 26 Nov 1996 14:47:12 -0800
I have some questions about basic asumtotics.  
1)  Is Convergence in Probability both necessary and sufficient
condition for consistency?  If not, example please.
2) To be a consitent estimator, does the Variance of it have to exist?
 [i.e.  Is Var(a_n) < infinity necessary condition for a_n to be a
consistent   estimator for a ? ]  
3) Could anyone provide INTUITIVE explanation for 
	1. Almost sure convergence => Convergence in Probability
	2. Convergence in Probability => Convergence in Distribution
Tatsuo Ochiai
tochiai@students.wisc.edu
Return to Top
Subject: Basic Asymtotics
From: Tatsuo Ochiai
Date: Tue, 26 Nov 1996 14:48:51 -0800
I have some questions about basic asumtotics.  
1)  Is Convergence in Probability both necessary and sufficient
condition for consistency?  If not, example please.
2) To be a consitent estimator, does the Variance of it have to exist?
 [i.e.  Is Var(a_n) < infinity necessary condition for a_n to be a
consistent   estimator for a ? ]  
3) Could anyone provide INTUITIVE explanation for 
        1. Almost sure convergence => Convergence in Probability
        2. Convergence in Probability => Convergence in Distribution
Thanks in advance.
Tatsuo Ochiai
tochiai@students.wisc.edu
Return to Top
Subject: LMedS regression
From: waynet@pendragon.cna.tek.com (Wayne Turner)
Date: 26 Nov 1996 13:29:56 -0800
I'm looking for algorithms to do least-median-square
regression.  Please email any references to published
algorithms or to software (commercial or public domain) and
I will post a summary.
Thanks,
Wayne
waynet@pendragon.cna.tek.com
Return to Top
Subject: Re: bootstrap methods
From: gregorich@psg.ucsf.edu (Steve Gregorich)
Date: 26 Nov 1996 21:48:45 GMT
Do a web search on "Resampling Stats"
There is a bunch of documents, books
even, that are available for download
from their web page.
HTH
Steve Gregorich
>Is anyone aware of tutorials, papers, 
>or other information which are accesible on the internet 
>(WWW).  Any help will be greatly appreciated.
Return to Top
Subject: Re: Power?
From: wpilib+@pitt.edu (Richard F Ulrich)
Date: 26 Nov 1996 22:18:36 GMT
I will not let certain mis-statements go unmentioned  --
Russ Boucher (at764@FreeNet.Carleton.CA) wrote:
<< ... deleted, citation;  and some comment, leading to a couple of 
   CORRECT statements, and then an error.  >>
RB: 	The problem with low power is that you run the risk of indecisive
: results, even if the phenomenon you're investigating is real. Stated
: differently, the effect may well be there, but without adequate power, you
: won't find it.
RB: 	Also, let's distinguish the difference between a Type I error and
: a Type II error. A Type I error is rejecting the null when it is true. A
: Type II error is failing to reject the null when it is true. Power is
: related to the latter, not the former.
  -- By ordinary and by mathematical definition, "failing to reject
the null when it is true" is NOT an error, assuming "it"  refers to 
the null.  "Power is related to the latter..."  reflects confusion
which is re-iterated even more clearly in the next Answer:
<< Q. deleted, about power greater of test is for p < .2 compared 
to p < .05>>
RB: 	Alpha is the probability of making a Type I error. Beta is the
: probability of making a Type II error. Thus, power is equal to 1 minus beta.
: In other words, power isn't really related to alpha at all.
  -- True, power= (1 minus beta), which is a SIMPLER formula than
the relation of power to alpha  --  but it is really WRONG to say 
that power isn't related to alpha at all.  For a given test, there 
is a strict relationship  among sample size and effect size, 
alpha and beta.  There are books that can show you tables, and 
computer programs that will plot you curves.  And there is an
interesting equivalence which exists by definition:  for the 
effect size nearing 0, for any N,  power = alpha.
Rich Ulrich, biostatistician              wpilib+@pitt.edu
Western Psychiatric Inst. and Clinic   Univ. of Pittsburgh
Return to Top
Subject: Re: Tricky sampling question
From: T.Moore@massey.ac.nz (Terry Moore)
Date: 26 Nov 1996 21:31:38 GMT
In article <961125095604.69@IASX02.LARC.NASA.GOV>, Ken Rutledge
 wrote:
> 
>   w.r.t. the following discussion "Tricky sampling question"
> 
>  wrote:
> 
> >My question is best presented by an example.  Say a runner wants to test
> >whether he can run faster with a particular pair of running shoes than
> >with another pair.  He does 10 runs with one pair of shoes and 10 runs
> >with the other pair.
The answer is clear. Use a paired design, one type of shoe on each
foot. If one is really better, the runner will run in circles :-) 
Terry Moore, Statistics Department, Massey University, New Zealand.
Imagine a person with a gift of ridicule [He might say] First that a
negative quantity has no logarithm; secondly that a negative quantity has
no square root; thirdly that the first non-existent is to the second as the
circumference of a circle is to the diameter. Augustus de Morgan
Return to Top
Subject: Capture and re-capture data
From: Chung Sai Ho
Date: Wed, 27 Nov 1996 11:59:10 +0800
Can anyone tell me where I can get some capture and recapture datasets?
I am interested in the relationship between growth and fish and want to study 
 various kinds of growth curve model. I prefer fish or any aquatic animal 
dataset with capture and recapture history (capture and recapture time), 
length and/or weight of the individual, age (estimated). More details are 
preferred.
Thank you very much!!
Return to Top
Subject: Re: Tricky sampling question
From: mdupin@kodak.com (Mike)
Date: Tue, 26 Nov 1996 11:30:02 -0600
Two references:
In Journal of Quality Technology, January 1996, p. 123, Lloyd Nelson has
an interesting article "Notes on the Use of Randomization in
Experimentation", and in that article he references an article by WJ
Youden that appeared in Technometrics February 1972, "Randomization and
Experimentation".
Quoting the abstract of Youden's article: "Randomization, often specified
as an indispensable requirement in experimental design, is required only
when the order or position of the experimental unit influences the order
of the unit."    
For your example, if you believe the day-to-day effects on running
performance (wind, etc.) occur independently, then you might argue that
there is no need to randomize, or that lack of randomization does not
render the t-test invalid.
My interpretation, anyway .....  
-----------------
> On Wed, 20 Nov 1996 13:25:27 -0500, "Michael Kleiman (SOC)"
>  wrote:
> 
> >My question is best presented by an example.  Say a runner wants to test
> >whether he can run faster with a particular pair of running shoes than
> >with another pair.  He does 10 runs with one pair of shoes and 10 runs
> >with the other pair.
  
 Instead the t-test is being performed on
> >the first 10 runs with each shoe, not on 10 runs randomly selected from
> >all runs.  The reason I think I'm still OK though is that while it's not
> >technically a random sample, everything that could affect the running
> >times other than the shoes is uncorrelated with the shoes.  For instance
> >the wind could be blowing harder for one of the runs, but this is
> >uncorrelated with the shoes because the wind doesn't decide to blow based
> >on my shoes.  If anyone could answer this question for me I'd greatly
> >appreciate it, the more so if you can also give me a reference I can 
> >consult. Of course, if doing a significance test in such a case is *not* 
> >OK then a lot of real-world testing is going to have to be trashed.  After
> >all, it'd be idiotic, not to mention impractical, for the runner to run
> >with the 2 shoes for the rest of his life and then randomly sample 10 runs 
> >from the resulting population!
Return to Top
Subject: I found out how to use fft data in MS excel
From: peter homel
Date: Tue, 26 Nov 1996 22:58:34 EST
This is a followup to a question I posted last night. I was trying to use the f
ft add on in Microsoft Excel and finding that the real and imaginary components
 that were returned by the fft were stored as a single array. Upon further
further exploration, there turn out to be a whole slew of engineering
functions in Excel that allow you to extract and work with complex number
arrays. Thanks to anyone who reads my original question and responds. In
the meantime, I think that I have enough information to go on my own.
PETER HOMEL PHD
HEALTH SCIENCE CENTER BROOKLYN
STATE UNIVERSITY OF NEW YORK
450 CLARKSON AVENUE BOX 7
BROOKLYN, NY 11203-2098
EMAIL: HOMEL@SACC.HSCBKLYN.EDU
       HOMEL@SNYBKSAC.BITNET
TEL: (718) 270-7424
FAX: (718) 270-7461
MOTTO: STATISTICS DON'T LIE!(PEOPLE DO!)
Return to Top
Subject: Statistics of outcomes of competitions
From: Eric Bartels
Date: 27 Nov 1996 09:25:50 +0000
Hello,
I am looking for pointers to publications in the biology/statistics
community where the following problem arose.   
        An experiment is conducted in which n individuals participate. Some
        individuals can compete with other individuals for a resource. For
        each individual the list of the possible opponents is fixed at the
        beginning. During the observation time some of these competitions take
        place and for each one the winner is noted. Some competitions may end
        in a tie in which case the competition is being treated as if it has
        not taken place.
        At the end we obtain for each individual the number of wins it has
        achieved during the observation time.
This outcome data is to be tested statistically in order to test
certain hypotheses about the population and certain subsets of it. 
I am curious to learn whether this or related problems have been dealt
with in the statistics / biology literature. Any paper where a similar
problem was analysed or pointers to the statistical method of choice
would be very much appreciated.
thanks,
Eric Bartels
Return to Top
Subject: Re: prediction contest
From: Gerrit.Jacobsen@hamburg.netsurf.de (Gerrit Jacobsen)
Date: Wed, 27 Nov 1996 09:06:59 GMT
cochen@iastate.edu (Cong Chen) wrote:
> I have overheard that there is an annual prediction contest. Please send me
>some information about that if you know it.  Thanks!!!
>P.S. I'm desperately trying to test my time series model. 
Perhaps you would like to share it with us ?
Gerrit Jacobsen
Return to Top

Downloaded by WWW Programs
Byron Palmer