Language:EN
Pages: 3
Words: 3578
Rating : ⭐⭐⭐⭐⭐
Price: $10.99
Page 1 Preview
the value the statistic based random sample size

The value the statistic based random sample size

  SAMPLING AND LARGE SAMPLE TESTS

 

For the purpose of determining population characteristics, instead of enumerating the entire population, the individuals in the sample only are observed. Then the sample characteristics are utilized to approximately determine or estimate the population. For example, on examining the sample of a particular stuff we arrive at a decision of purchasing or rejecting that stuff. The error involved in such approximation is known as sampling error and is inherent and unavoidable in any and every sampling scheme. But sampling results in considerable gains, especially in time and cost not only in respect of making observations of characteristics but also in the subsequent handling of the data.

Sampling is quite often used in our day – to – day practical life. For example, in a shop we assess the quality of sugar, wheat or any other commodity by taking a handful of it from the bag and then decide to purchase it or not. A housewife normally tests the cooked products to find if they are properly cooked and contain the proper quantity of salt.

(ii)        Random sampling

(iii)       Stratified sampling

Purposive sampling is one in which the sample units are selected with definite purpose in view. For example, if we want to give the picture that the standard of living has increased in the city of New Delhi, we may take individuals in the sample from rich and posh localities like Defence Colony, South Extension, Golf Links, Jor Bagh, Chanakyapuri, Greater Kailash etc. and ignore the localities where low income group and the middle class families live. This sampling suffers from the drawback of favouritism and nepotism and does not give a representative sample of the population.

10.2.2       Random sampling:

Suppose we want to select ‘r’ candidates out of n. we assign the numbers one to n, one number to each candidate and write these numbers on n slips which are made as homogeneous as possible in shape, size, etc. these slips are then put in a bag and thoroughly shuffled and then ‘r’ slips are drawn one by one. The ‘r’ candidates corresponding to the numbers on the slips drawn will constitute the random sample.

 

Here the entire heterogeneous population is divided into a number of homogeneous groups, usually termed as strata, which differ from one another but each of these groups is homogenous within itself. Then units are sampled at random from each of this stratum, the stratum in the population. The sample, which is the aggregate of the sampled units of each of the stratum, is termed as stratified sample and the technique of drawing this sample is known as stratified sampling. Such a sample is by far the best and can safely be considered as representative of the population from which it has been drawn.

 

10.3.1       Sampling distribution of a statistic:

If we draw a sample of size n from a given finite population of size N, then the total number of possible samples is:

Sample number t s2
1 t1 s12
2 t2 s22
3 t3 S32
. . .
. . . .
k tk sk2

 

 

 

10.4 TESTS OF SIGNIFICANCE:

Since, for large n, almost all the distributions, eg., Binomial, Poisson, Negative binomial, Hyper geometric, t, F, chi square can be approximated very closely by a normal probability curve, we use the normal test of significance for large samples. Some of the well known tests of significance for studying such differences for small samples are t-test, F-test and Fisher’s z –transformation.

 

 

10.5.1       Alternative Hypothesis:

The alternative hypothesis in (i) is known as a two tailed alternative and the alternative in (ii) and (iii) are known as right tailed and left-tailed alternatives respectively. The setting of alternative hypothesis is very important since it enables us to decide whether we have to use as single-tailed or two tailed test.

 

Type II Error: Accept H0 when it is wrong, that is accept H0 when H1 is true.

If we write.

Thus P { Reject a lot when it is good} = α

and P { Accept a lot when it is bad } = β

P( t Є ω| H0 ) = α,

P(t Є | H1 ) = β

10.7.1       One tailed and two tailed tests:

In any test, the critical region is represented by a portion of the area under the probability curve of the sampling distribution of the test statistic.

Is a single tailed test. In the right test (H1: μ > μ0), the critical region lies entirely in the right tail of the sampling distribution or , while for the left tail test (H1: μ < μ0), the critical region is entirely in the left tail or the distribution.

A test of statistical hypothesis where the alternative hypothesis is two tailed such as:

The value of test statistic which separates the critical region and the acceptance region is called the critical value or significant value. It depends upon:

(i)          The level of significance used, and

P(|Z|> Zα) = α ------------------(1)

That is Zα is the value so that the total area of the critical region on both tails is α. Since normal probability curve is a symmetrical curve, from (1), we get

That is the area of each tail is α/2. Thus Zα is the value such that area to the right of Zα is α/2 and to the left of - Zα is α/2.

In case of single tail alternative, the critical value Zα is determined so that total area to the right of it is α and for left tailed test the total area to the left of - Zα is α.

1.          Null hypothesis: set up the null hypothesis H0.

2.          Alternative hypothesis: set up the alternative hypothesis H1. This will enable us to decide whether we have to use a single tailed test or two tailed test.

5.          Conclusion: we compare z the computed value of z in step 4 with the significant value zα, at the given level of significance, ‘α’.

If |Z| < zα, that is if the calculated value of Z is less than zα we say it is not significant. By this we mean that the difference t – E(t) is just due to  fluctuations of sampling and the sample data do not provide us sufficient evidence against the null hypothesis which may therefore, be accepted.

If X ≈ n( μ, σ2), then z =

Thus from the normal probability tables, we have

P ( -1.96 ≤ Z ≤ 1.96 ) = 0.95 that is P(|Z| ≤ 1.96) = 0.95

è  P(|Z| > 1.96 ) = 1 – 0.95 = 0.05

(i)          Compute the test statistic Z under H0.

(ii)        If |Z| >3, H0 is always rejected.

= 0.5 – 0.45

= 0.05

 

10.9 SAMPLING OF ATTRIBUTES:

E(X) = nP and V(X) = n PQ

Where Q = 1- P, is the probability of failure.

10.9.2       Test of significance for difference of proportions:

Suppose we want to compare two distinct populations with respect to the prevalence of a certain attribute, say A, among their members. Let X1, X2 be the number of persons possessing the given attributes respectively. Then sample proportions are given by

Since for large sample, p1 and p­are asymptotically normally distributed (p1- p2) is also normally distributed. Then the standard variable corresponding to the difference (p1- p2) is given by

10.10       SAMPLING OF VARIABLES:

In the case of sampling variables each member of the population provides the value of the variable and the aggregate of these values forms the frequency distribution of the population. From the population, a random sample of size n can be drawn by any of the sampling methods discussed before which is same as choosing n values of the given variables from the distribution.

Now E () = 

Since xi is a sample observation from the population Xi, (I = 1, 2, ……N) it can take any one of the values X1, X2,……………. XN each with equal probability 1/N.

----------------------------(1)

We have V(xi) = E[xi-E(xi)]2 = E(xi-μ]2

In particular,

-----------------------(2)

-------------(4)

Since  , sample variance is not an unbiased estimate of population variance.

Therefore S2 is an unbiased estimate of the population variance σ2.

 

Let x1, x2,…… xn be a random sample of size n from a population with variance σ2, then the sample mean is given by

Standard error 

Under the null hypothesis, H0that the sample has been drawn from a population with mean μ and variance σ2 that is there is no significant difference between the sample mean and population mean (μ), the test statistic is:

 and 

Also , being the difference of two independent normal variates is also a normal variate. The Z standard normal variable corresponding to is given by

The covariance term vanishes, since the sample mean and are independent.

Thus under H0: μ1 = μ2, the test statistic becomes

for large samples.

But in case of large samples, the standard error(S.E.) of the difference of the sample standard deviations is given by

 

 

You are viewing 1/3rd of the document.Purchase the document to get full access instantly

Immediately available after payment
Both online and downloadable
No strings attached
How It Works
Login account
Login Your Account
Place in cart
Add to Cart
send in the money
Make payment
Document download
Download File
img

Uploaded by : Miss Carly Conway

PageId: DOCD65E7E0