Professional Documents
Culture Documents
+ =
|
|
.
|
\
|
=
2
1
2
1
1 1
1
1
1 1 1
W n N
W n N
nW n
E D ;
( )( )
( )
(
+ =
|
|
.
|
\
|
=
2
2
2
2
2 2
2
1
1 1 1
W n N
W n N
nW n
E D
N
Y N Y N
Y
2 2 1 1
+
= ;
2 2 1 1
N
X N X N
X
+
=
ASSUMPTIONS
Consider following in light of Figure 1 before formulating
an imputation based estimation procedure:
1) The sample of size n is drawn by SRSWOR and post-
stratified into two groups of size n
1
and
n
2
(n
1
+ n
2
= n)
according to R and NR group respectively
2) The information about Y variable in sample is
completely available.
3) The sample means of both groups
1
y and
2
y are
known such that
2 2 1 1
n
y n y n
y
+
=
which is sample mean on n units.
4) The population means
1
X and X are known.
5) The population size N and sample size n are known.
Also, N
1
and
N
2
are known by past data, past experience
or by guess of the investigator (N
1
+ N
2
= N).
6) The sample mean of auxiliary information
1
x is only
known for R-Group, but information about
2
x of NR-group
is missing. Therefore
2 2 1 1
n
x n x n
x
+
=
could not be obtained due to absence of
2
x .
7) Other population parameters are assumed known, in
either exact or in ratio from except the Y , 1 Y and 2 Y .
PROPOSED CLASS OF ESTIMATION STRATEGY
To estimate population meanY , in setup of Figure 1, a
problem to face is of missing observations related to
2
x ,
therefore, usual ratio, product and regression estimators
are not applicable. Singh and Shukla (1987) have
proposed a factor type estimator for estimating population
mean Y . Shukla et al. (1991), Singh and Shukla (1993)
and Shukla (2002) have also discussed properties of
factor-type estimators applicable for estimating
population mean. But all these cannot be useful due to
unknown information
2
x
. In order to solve this, an
imputation
( )
4
*
2
x
is adopted as:
( )
4 2
*
x =
( ) { }
(
+
n N
X f X f n X N 2
1
1
(1)
80 Afr. J. Math. Comput. Sci. Res.
Population (N=N
1
+ N
2
)
R-group NR-group
N
1
1 Y 1 X N
2
2 Y
2 X
2
1Y
S
2
1X
S
2
2Y
S
2
2X
S
Sample (n =n
1
+ n
2
)
R-group NR-group
n
1
1 y n
2
1 x 2 y 2 x
Fig : 1.1
2.0 ASSUMPTIONS :
Consider following in light of figure
Figure 1. Setup to estimate population mean Y .
The logic for this imputation is to utilize the non-sampled
part of the population of X for obtaining an estimate of
missing
2
x and generate
( ) 2
x for x as describe as
follows:
And
( ) 4
x =
( )
(
(
+
+
2 1
4
2
2 1 1
N N
x N x N
*
(2)
The proposed class of imputed factor-type estimator is:
[( y
FT
)
D
]
k
=
( )
( )
( )
( )
(
(
+ +
+ +
|
|
.
|
\
|
+
4
4
2 2 1 1
x C X fB A
x fB X C A
N
y N y N
(3)
Where < < k 0 and k is a constant and
A = (k 1) (k 2 ); B = (k 1) (k 4); C = (k 2) (k 3) (k 4); f = n / N.
LARGE SAMPLE APPROXIMATION
Consider the following for large n:
( )
( )
( )
( )
+ =
+ =
+ =
+ =
4 2 2
3 1 1
2 2 2
1 1 1
1
1
1
1
e X x
e X x
e Y y
e Y y
(4)
where,
1
e ,
2
e ,
3
e and
4
e are very small numbers 0 >
i
e
(i = 1,2,3,4).
Using the basic concept of SRSWOR and the concept of
post-stratification of the sample n into n
1
and n
2
(Cochran,
2005; Hansen et al., 1993; Sukhatme et al., 1984; Singh
and Choudhary, 1986; Murthy, 1976), we get
( ) ( ) | |
( ) ( ) | |
( ) ( ) | |
( ) ( ) | |
= =
= =
= =
= =
0 E E E
0 E E E
0 E E E
0 E E E
2 4 4
1 3 3
2 2 2
1 1 1
n e e
n e e
n e e
n e e
(5)
Assume the independence of R-group and NR-group
representation in the sample, the following expression
could be obtained:
| | ( ) | |
1
2
1
2
1
E E E n e e =
(
(
)
`
|
|
.
|
\
|
=
1
2
1
1
1 1
E n C
N n
Y
(
(
)
`
|
|
.
|
\
|
=
2
1
1
1 1
E
Y
C
N n
(
|
.
|
\
|
=
2
1 1
1
Y
C
N
D
(6)
| | ( ) | |
2
2
2
2
2
e E E n E e =
(
(
)
`
|
|
.
|
\
|
=
2
2
2
2
1 1
E n C
N n
Y
(
|
.
|
\
|
=
2
2 2
1
Y
C
N
D
(7)
| | ( ) | |
1
2
3
2
3
E E E n e e =
(
|
.
|
\
|
=
2
1 1
1
X
C
N
D
(8)
and
| |
(
|
.
|
\
|
=
2
2 2
2
4
1
E
X
C
N
D e
(9)
| | ( ) | |
1 3 1 3 1
E E E n e e e e =
(
(
)
`
|
|
.
|
\
|
=
1 1 1 1
1
1 1
E n C C
N n
X Y
(
|
.
|
\
|
=
X Y
C C
N
D
1 1 1 1
1
(10)
| | ( ) | | 0 E E E
2 1 2 1 2 1
= = n , n e e e e
(11)
| | 0 E
4 1
= e e
(12)
| | ( ) | | 0 , E E E
2 1 3 2 3 2
= = n n e e e e
(13)
| |
X Y
C C
N
D e e
2 2 2 2 4 2
1
E
|
.
|
\
|
=
(14)
| | 0 E
4 3
= e e
(15)
The expressions (11), (12), (13) and (15) are true under
the assumption of independent representation of R-group
and NR-group units in the sample. This is introduced to
simplify mathematical expressions.
Shukla et al. 81
Theorem 1
The estimator [( y
FT
)
D
]
k
could be expressed under large
sample approximation in the form:
[( y
FT
)
D
]
k
= o
3
Y [1 + s
1
W
1
e
1
+ s
2
W
2
e
2
][1 + (o
3
|
3
)e
3
(o
3
|
3
)|
3
2
3
e + (o
3
|
3
)
2
3
|
3
3
e
..
]
Proof
Rewrite
( ) 4
x as in Equation 2:
( ) 4
x =
( )
(
(
+
+
2 1
4
*
2
2 1 1
N N
x N x N
Where,
( )
4 2
*
x =
( ) { }
(
+
n N
X f X f n X N 2
1
1
) 4 (
x
=
( )
{ }
(
(
)
`
+
+ +
n N
X f X f n X N
N e X N
N
2 1
2 3
1
1
) 1 (
1
1
=
( ) ( ) ( ) | |
N
X f X f n X N p e X N 2 1
3
1
1
1 1 + + +
=
( )
(
+ +
N
e X N X f pn X pnf X pN X N
3
1
1
2 1 1
1
1
=
( ) ( )
(
+ +
N
e X N X f pn X pnf N X pN
3
1
1
2 1
1
1
= ( ) | |
3 1 1 2 1
2
1
1 e r W r ) f ( pf r pf W p X + +
= | |
3 1 1
e r W X +
(16)
Where,
p =
n N
N
2
; = p + (W
1
- pf
2
)r
1
pf (1-f ) r
2
.
Now, the estimator [( y
FT
)
D
]
k
under approximation and
using (16) is
[( y
FT
)
D
]
k
=
( )
( )
( )
( )
(
(
+ +
+ +
|
|
.
|
\
| +
4
4
2 2 1 1
x C X fB A
x fB X C A
N
y N y N
=
( ) ( ) ( ) ( )
( ) ( )
(
+ + +
+ + +
(
+ + +
X e r W C X fB A
X e r W fB X C A
N
e Y N e Y N
3 1 1
3 1 1 2
2
2 1
1
1
1 1
=
| |
( )
( )
(
+ + +
+ + +
+ +
3 1 1
3 1 1
2 2 2 1 1 1
1
e r CW C fB A
e r fBW C fB A
e W s e W s Y
= | |
(
+
+
+ +
3 4 3
3 2 1
2 2 2 1 1 1
1
e
e
e W s e W s Y
= o3 | |
2 2 2 1 1 1
1 e W s e W s Y + + (1 + o3e3)(1 + |3e3)
-1
82 Afr. J. Math. Comput. Sci. Res.
Where,
1
= A + fB + C;
2
= fBW
1
r
1;
3
= A + fB + C;
4
= CW
1
r
1
;
r
1
=
X
X
1
; r
2
=
X
X
2
; s
1
=
Y
Y
1
; s
2
=
Y
Y
2
; o
3
=
1
2
;
|
3
=
3
4
;
o
3
=
3
1
.
We can further express the above into following:
= o
3Y [1 + s
1
W
1
e
1
+ s
2
W
2
e
2
] (1 + o
3
e
3
) (1 |
3
e
3
+
2
3
|
2
3
e
3
3
|
3
3
e
.
)
=o
3
Y [1+s
1
W
1
e
1
+s
2
W
2
e
2
][1+(o
3
|
3
)e
3
(o
3
|
3
)|
3
2
3
e
+(o
3
|
3
)
2
3
|
3
3
e .]
BIAS AND MEAN SQUARED ERROR
Define E(.) for expectation, B(.) for bias and M(.) for
mean squared error, then the first order of
approximations could be established for i, j = 1, 2, 3, .....
as
| | 0 E
2 1
=
j i
e e when i+j > 2
| | 0 E
3 1
=
j i
e e when i+j > 2
| | 0 E
3 2
=
j i
e e when i+j > 2
(17)
Theorem 2
The [( y
FT
)
D
]
k
is a biased estimator of Y with the amount
of bias to the first order of approximation:
( ) | | ( ) ( ) { }
(
|
.
|
\
|
=
Y X X k D FT
C W s C
N
D C Y y
1 1 1 1 1 3 1 3 3 1 3 3
1
1 B | | o o o
Proof
( ) | | ( ) { } | | Y y y
k D FT
k D FT
= E B
Using theorem 1 and taking expectations
E[( y
FT
)
D
]
k
= o
3
Y E[1 + (o
3
|
3
)e
3
(o
3
|
3
) |
3
2
3
e
+
s
1
W
1
e
1
{1 + (o
3
|
3
) e
3
(o
3
|
3
)
|
3
2
3
e }
+ s
2
W
2
e
2
{1+ (o
3
|
3
) e
3
(o
3
|
3
) |
3
2
3
e }]
= o
3
Y [1
(o
3
|
3
)|
3
E(
2
3
e )+ s
1
W
1
(o
3
|
3
) E(e
1
e
3
) + s
2
W
2
(o
3
|
3
)E(e
2
e
3
)]
= ( )
|
.
|
\
|
2
1 1 3 3 3 3
1
1
X
C
N
D Y | | o o ( )
(
(
|
.
|
\
|
+
X Y
C C
N
D W s
1 1 1 1 1 1 3 3
1
| o
= ( ) { }
(
|
.
|
\
|
Y X X
C W s C C
N
D Y
1 1 1 1 1 3 1 1 3 3 3
1
1 | | o o
Therefore,
( ) | | ( ) { } | | Y y y
k D FT
k D FT
= E B
= ( ) ( ) { }
(
|
.
|
\
|
Y X X
C W s C
N
D C Y
1 1 1 1 1 3 1 3 3 1 3 3
1
1 | | o o o
Theorem 3
The mean squared error of [( y
FT
)
D
]
k
is
M [( y
FT
)
D
]
k
= ( ) { }
(
(
|
.
|
\
|
+
+ + |
.
|
\
|
+
2
2
2
2
2
2
2
3 2 1 1 1 1 3
2
1 2
2
1
2
1 1 1
2
3
2 1
2
1
1
Y X Y X Y
C W s
N
D C C s J C J C s J
N
D Y o o
Where
2
1
2
3 1
W J o = , ( ) ( ) ( ) { }
3 3 3 3 3 3 3 3 2
1 2 | o | o o | o o = J ; ) )( 1 2 (
3 3 3 3 1 3
| o o o =W J
Proof
M [( y
FT
)
D
]
k
= E[{( y
FT
)
D
}
k
Y ]
2
Using Theorem 1, we can express
M[( y
FT
)
D
]
k
= E[ o
3
Y {1 + s
1
W
1
e
1
+ s
2
W
2
e
2
}{1+ (o
3
|
3
)e
3
(o
3
|
3
)|
3
2
3
e + (o
3
|
3
)
2
3
|
3
3
e
...}Y ]
2
Using large sample approximations of (17) we could
express
=
2
Y E [(o
3
1)+o
3
{(o
3
|
3
)e
3
(o
3
|
3
)|
3
2
3
e +(s
1
W
1
e
1
+ s
2
W
2
e
2
)+ (o
3
|
3
)(s
1
W
1
e
1
+ s
2
W
2
e
2
)
e
3
}]
2
=
2
Y E [(o
3
1)
2
+
2
3
o {(o
3
|
3
)
2 2
3
e
+
2
1
s
2
1
W
2
1
e
+
2
2
s
2
2
W
2
2
e +2s
1
s
2
W
1
W
2
e
1
e
2
+ 2(o
3
|
3
)(s
1
W
1
e
1
e
3
+ s
2
W
2
e
2
e
3
)}+ 2o
3
(o
3
1) {(o
3
|
3
)e
3
(o
3
|
3
)|
3
2
3
e
+ (s
1
W
1
e
1
+ s
2
W
2
e
2
) + (o
3
|
3
) (s
1
W
1
e
1
e
3
+ s
2
W
2
e
2
e
3
)}]
Using (5), (11) and (12) we rewrite,
=
2
Y [ (o31)
2
+
2
3
o {(o3 |3)
2
E(
2
3
e )
+
2
1
s
2
1
W
E(
2
1
e )
+
2
2
s
2
2
W E(
2
2
e )
+2(o3 |3) s1W1 E(e1 e3)}
+ 2o3(o3 1){ (o3|3)|3 E(
2
3
e
)+ (o3 |3)s1W1 E(e1e3 )}]
( ) ( )
|
.
|
\
|
+ |
.
|
\
|
+ =
2
1 1
2
1
2
1
2
1 1
2
3 3
2
3
2
3
2 1 1
1
Y X
C
N
D W s C
N
D Y | o o o
( )
)
`
|
.
|
\
|
+ |
.
|
\
|
+
X Y Y
C C
N
D W s C
N
D W s
1 1 1 1 1 1 3 3
2
2 2
2
2
2
2
1
2
1
| o
( ) ( )
)
`
|
.
|
\
|
+
2
1 1 3 3 3 3 3
1
1 2
X
C
N
D | | o o o ( )
(
(
|
.
|
\
|
+
X Y
C C
N
D W s
1 1 1 1 1 1 3 3
1
| o
( ) {
|
.
|
\
|
+ =
2
1
2
1
2
1
2
3 1
2
3
2 1
1
Y
C W s
N
D Y o o ( ) ( ) ( ) { } 1 2
2
1 3 3 3 3 3 3 3 3 X
C | o | o o | o o +
( )( ) }
X Y
C C W s
1 1 1 1 1 3 3 3 3
1 2 | o o o +
(
(
|
.
|
\
|
+
2
2
2
2
2
2
2
3 2
1
Y
C W s
N
D o
( ) {
+ |
.
|
\
|
+ =
2
1 2
2
1
2
1 1 1
2
3
2 1
1
X Y
C J C s J
N
D Y o }
(
(
|
.
|
\
|
+ +
2
2
2
2
2
2
2
3 2 1 1 1 1 3
1
2
Y X Y
C W s
N
D C C s J o
SOME SPECIAL CASES
The term A, B and C are functions of k. In particular,
there are some special cases:
Case 1
When k = 1
A = 0; B = 0; C = 6;
1
= -6;
2
= 0;
3
= - 6;
4
= -6r
1
w
1;
o
3
= 0; |
3
=
1 1
W r
; o
3
=
1
;
J
1
=
2
2
1
W
; J
2
=
4
2
1
2
1
2 3
) ( W r
; J
3
=
3
2
1 1
2
) ( W r
;
The estimator [( y
FT
)
D
]
k
along with bias and m.s.e. under
case i is:
[( y
FT
)
D
]
k =1
=
(
(
(
+
) (
x
X
N
y N y N
4
2 2 1 1
(18)
B[( y
FT
)
D
]
k =1
=Y
-3
[(1- )
2
+
|
.
|
\
|
N
D
1
1
r
1
2
1
W C
1X
{r
1
C
1X
- s
1
1
C
1Y
}]
(19)
M[( y
FT
)
D
]
k=1
=
2
Y
-4
[(1 )
2
2
+
2
1
W
|
.
|
\
|
N
D
1
1
{
2
2
1
s
2
1Y
C + (3 - 2 )
2
1
r
2
1X
C
+ 2(
- 2) r
1
s
1
1
C
1Y
C
1X
}+
|
.
|
\
|
N
D
1
2
2 2
2
W
2
2
s
2
2Y
C ]
(20)
Case 2
When k = 2
A = 0; B = 2; C = 0;
1
= -2f ;
2
= -2fW
1
r
1
;
3
= -2f;
4
= 0;
o
3
=
1
1 1
W r ; |
3
= 0; o
3
= ; J
1
=
2 2
1
W ; J
2
=
2
1
2
1
W r ; J
3
= ) ( W r 1 2
2
1 1
;
[( y
FT
)
D
]
k =2
=
(
(
(
(
+
X
x
N
y N y N
) ( 4
2 2 1 1
(21)
B [( y
FT
)
D
]
k =2
=
(
|
.
|
\
|
+
X Y
C C s r W
N
D ) ( Y
1 1 1 1 1
2
1 1
1
1
(22)
M [( y
FT
)
D
]
k =2
=
2
Y [( 1)
2
+
2
1
W
|
.
|
\
|
N
D
1
1
{
2
2
1
s
2
1Y
C + r
1
2 2
1X
C + 2(2 1)s
1
r
1
1
C
1Y
C
1X
}
+
|
.
|
\
|
N
D
1
2
2 2
2
W
2
2
s
2
2Y
C ]
(23)
Shukla et al. 83
Case 3
When k = 3
A = 2; B = 2; C = 0; 1= 2(1- f); 2 = -2fW1r1; 3= 2(1-f ); 4=0; o3 =
f
r fW
1
1 1
|3 = 0; o3 =
f
f
1
1
; J1=
( )
( )
2
2
1
2
1
1
f
W f
; J2 =
( )
2
2
1
2
1
2
1 f
r W f
; J3 =
{ }
( )
2
1
2
1
1
1 2
f
f f r fW
;
[( y
FT
)
D
]
k =3
=
(
(
+
|
|
.
|
\
| +
X ) f (
x f X
N
y N y N
) (
1
4
2 2 1 1
(24)
B [( y
FT
)
D
]
k =3
=
(
|
.
|
\
|
X Y
C C s r W
N
D ) ( ) f ( f Y
1 1 1 1 1
2
1 1
1
1
1 1
(25)
M [( y
FT
)
D
]
k =3
=
2
Y (1 f )
-2
[f
2
(1 )
2
+
|
.
|
\
|
N
D
1
1
2
1
W {(1f)
2 2
1
s
2
1Y
C + f
2 2
1
r
2
1X
C
+ 2(2f f 1)fr
1
s
1
1
C
1Y
C
1X
} +
|
.
|
\
|
N
D
1
2
(1f)
2 2
2
W
2
2
s
2
2Y
C ]
(26)
Case 4
When k = 4;
A = 6; B = 0; C = 0 ;
1
= 6;
2
= 0;
3
= 6;
4
= 0; o
3
= 0; |
3
= 0; o
3
=
1; J
1
=
2
1
W ; J
2
= 0; J
3
= 0;
[( y
FT
)
D
]
k =4
=
(
(
+
N
y N y N
2 2 1 1
(27)
B[( y
FT
)
D
]
k =4
= 0
(28)
V[( y
FT
)
D
]
k =4
=
2
2
2
2
2
2 2
2
1
2
1
2
1 1
1 1
Y Y
C Y W
N
D C Y W
N
D |
.
|
\
|
+ |
.
|
\
|
(29)
ESTIMATOR WITHOUT IMPUTATION
Throughout the discussion, the assumption is unknown
value of
2
x . This is imputed by the term ( )
4 2
*
x , to provide
the generation of
( ) 4
x [Equations (1) and (2)]. Suppose
the
2
x is known, then there is no need of imputation and
the proposed (2) and (3) reduces into:
|
.
|
\
| +
=
N
x N x N
x
2 2 1 1 (*)
(30)
84 Afr. J. Math. Comput. Sci. Res.
( ) | |
( ) |
|
.
|
\
|
+ +
+ +
|
.
|
\
| +
=
*
(*)
k w FT
x C X ) fB A (
x fB X ) C A (
N
y N y N
y
2 2 1 1
(31)
Where, k is a constant , 0 < k < and
A = (k 1) (k 2); B = (k 1) (k 4); C = (k 2) (k 3) (k 4); f = n/N.
Theorem 4
The estimator ( ) | |
k w FT
y is biased for Y with the amount
of bias
( ) | | ( ) { }
|
.
|
\
|
=
X
'
Y X
' '
k w FT
C r C s C r W
N
D y B
1 1 2 1 1 1 1 1
2
1 1 2 1
1
{ }
(
(
|
.
|
\
|
+
X
'
Y X
C r C s C r W
N
D
2 2 2 2 2 2 2 2
2
2 2
1
Where,
( ) C fB A / fB
'
+ + =
1
; ( ) C fB A / C
'
+ + =
2
.
Proof
The estimator ( ) | |
k w FT
y could be approximate like:
( ) | |
( )
( )
( )
( ) |
|
.
|
\
|
+ +
+ +
|
.
|
\
| +
=
*
*
k w FT
x C X fB A
x fB X C A
N
y N y N
y
2 2 1 1
( ) ( )
(
+ + +
=
N
e Y N e Y N
2 2 2 1 1 1
1 1
( ) ( ) ( ) { }
( ) ( ) ( ) { }
(
+ + + + +
+ + + + +
4 2 2 3 1 1
4 2 2 3 1 1
1 1
1 1
e X N e X N C X fB A N
e X N e X N fB X C A N
| |
( ) ( )
( ) ( )
(
+ + + +
+ + + +
+ + =
4 2 2 3 1 1
4 2 2 3 1 1
2 2 2 1 1 1
e r W e r W C C fB A
e r W e r W fB C fB A
e Y W e Y W Y
| | ( ) | |
4 2 2 3 1 1 1 2 2 2 1 1 1
1 e r W e r W e Y W e Y W Y
'
+ + + + = ( ) | |
1
4 2 2 3 1 1 2
1
+ + e r W e r W
'
Expanding above using binominal expansion, and
ignoring ( )
l
j
k
i
e e terms for ( k + l) > 2, (k, l = 0, 1, 2 ........),
(i, j = 1, 2, 3, 4 ); the estimator result into
( ) | | ( ) ( ) ( )
2 1 2 2 2 2 1 1 1 1 2 1
1 1 A A A A A A + + + + + = e Y W e Y W Y Y y
k w FT
(32)
Where,
A
1
= ( ) ( )
4 2 2 3 1 1 2 1
e r W e r W
' '
+ ; A
2
= ( ) ( )
2
4 2 2 3 1 1 2 1 2
e r W e r W
' ' '
+
And
W
1
r
1
+W
2
r
2
= 1 holds.
Further, one can derive up to first order of approximation
according to the following;
(i) E [A
1
] = 0
(ii) E| |
2
1
A = ( )
(
|
.
|
\
|
+ |
.
|
\
|
2
2 2
2
2
2
2
2
1 1
2
1
2
1
2
2 1
1 1
X X
' '
C
N
D r W C
N
D r W
(iii) E| |
2
A = ( )
(
|
.
|
\
|
+ |
.
|
\
|
2
2 2
2
2
2
2
2
1 1
2
1
2
1 2 1 2
1 1
X X
' ' '
C
N
D r W C
N
D r W
(iv) E| |
1 1
A e = ( )
Y X
' '
C C
N
D r W
1 1 1 1 1 1 2 1
1
|
.
|
\
|
(v) E| |
1 2
A e = ( )
Y X
' '
C C
N
D r W
2 2 2 2 2 2 2 1
1
|
.
|
\
|
(vi) E| |
2 1
A e = ( ) | |
1
0 under 0
n
(vii) E| |
2 2
A e = ( ) | |
1
0 under 0
n
The bias of estimator without imputation is
( ) | | ( ) { } | | Y y y
k w FT
k w FT
= E B
| | ) 1 ( ) 1 ( ) (
2 1 2 2 2 2 1 1 1 1 2 1
A A + + A A + + A A = e Y W e Y W Y E
( ) ( ) | |
2 1 2 2 2 1 1 1 1
A A A E Y ) e ( E Y W e E Y W + =
( ) { }
|
.
|
\
|
=
X
'
Y X
' '
C r Y C Y C r W
N
D
1 1 1 1 1 1 1 1
2
1 1 2 1
1
{ }
(
(
|
.
|
\
|
+
X
'
Y X
C r Y C Y C r W
N
D
2 2 1 2 2 2 2 2
2
2 2
1
( ) { }
|
.
|
\
|
=
X
'
Y X
' '
C r C s C r W
N
D Y
2 1 2 1 1 1 1 1
2
1 1 2 1
1
{ }
(
(
|
.
|
\
|
+
X
'
Y X
C r C s C r W
N
D
2 2 2 2 2 2 2 2
2
2 2
1
Theorem 5
The mean squared error of the estimator ( ) | |
k w FT
y is
( ) | | { }
(
+ + |
.
|
\
|
=
X Y X Y
k w FT
C C T T C T C T W
N
D Y y
1 1 1 2 1
2
1
2
2
2
1
2
1
2
1 1
2
2
1
M
{ }
(
(
+ + |
.
|
\
|
+
X Y X Y
C C S S C S C S W
N
D
2 2 2 2 1
2
2
2
2
2
2
2
1
2
2 2
2
1
Where,
T
1
= s
1;
T
2
= ( )
1 2 1
r
' '
; S
1
= s
2
; S
2
= ( )
2 2 1
r
' '
;
Proof
( ) | | ( ) { } | |
2
E M Y y y
k w FT k w FT
=
[( y
FT
)
w
]
k
( ) ( ) ( ) | |
2
2 1 2 2 2 2 1 1 1 1 2 1
1 1 E A A A A A A + + + + = e Y W e Y W Y
( ) ( ) ( ) ( )
1 1 1 1
2
2
2
2
2
2
2
1
2
1
2
1
2
1
2
E 2 E E E A A e Y Y W e Y W e Y W Y + + + =
( ) ( )
2 1 2 1 2 1 1 2 2 2
E 2 E 2 e e Y Y W W e Y Y W + + A
( )
)
`
|
.
|
\
|
+ |
.
|
\
|
=
2
2 2
2
2
2
2
2
1 1
2
1
2
1 2 1
2
1 1
2
X X
' '
C
N
D r W C
N
D r W Y
1 1
2
2 2
2
2
2
2
2
1 1
2
1
2
1
)
`
|
.
|
\
|
+ |
.
|
\
|
+
Y Y
C
N
D s W C
N
D s W
( )
)
`
|
.
|
\
|
+
X Y
' '
C C
N
D r W s W
1 1 1 1 1 1 2 1 1 1
1
2
( )
(
(
)
`
|
.
|
\
|
+
X Y
' '
C C
N
D r W s W
2 2 2 2 2 2 2 1 2 2
1
2
{ }
(
+ + |
.
|
\
|
=
X Y X Y
C C T T C T C T W
N
D Y
1 1 1 2 1
2
1
2
2
2
1
2
1
2
1 1
2
2
1
{ }
(
(
+ + |
.
|
\
|
+
X Y X Y
C C S S C S C S W
N
D
2 2 2 2 1
2
2
2
2
2
2
2
1
2
2 2
2
1
Remark 1
At k = 1, k = 2, k = 3 and k = 4, there are some special
cases of non-imputed estimators with the respective bias
and mean squared error as laid down with the following
Case 5
When k = 1
A = 0 ; B = 0 ; C = 6 ;
'
1
= 0;
'
2
= 1; T
1
= s
1
; T
2
= -r
1;
S
1
= s
1
; S
2
=
r
2;
;
A = 0 ; B = 0 ; C = 6 ;
'
1
= 0;
'
2
= 1; T
1
= s
1
; T
2
= -r
1;
S
1
= s
1
; S
2
=
r
2;
A = 0 ; B = 0 ; C = 6 ;
'
1
= 0;
'
2
= 1; T
1
= s
1
; T
2
= -r
1;
S
1
= s
1
; S
2
=
r
2;
( ) | |
( ) |
|
.
|
\
|
|
.
|
\
| +
=
= * k w FT
x
X
N
y N y N
y
2 2 1 1
1
( ) | | { }
|
.
|
\
|
=
= X Y X k w FT
C r C s C r W
N
D Y y
1 1 1 1 1 1 1
2
1 1 1
1
B { }
(
(
|
.
|
\
|
+
X Y X
C r C s C r W
N
D
2 2 2 2 2 2 2
2
2 2
1
( ) | | { }
+ |
.
|
\
|
=
= X Y X Y k w FT
C C r s C r C s W
N
D Y y
1 1 1 1 1
2
1
2
1
2
1
2
1
2
1 1
2
1
2
1
M
{ }
(
(
+ |
.
|
\
|
+
X Y X Y
C C r s C r C s W
N
D
2 2 2 2 2
2
2
2
2
2
2
2
2
2
2 2
2
1
Case 6
When k = 2
A = 0; B = 2; C = 0;
'
1
= 1;
'
2
= 0; T1 = s1; T2 = r1; S1= s2; S2 = r2.
Shukla et al. 85
( ) | |
( )
|
|
.
|
\
|
|
.
|
\
| +
=
=
X
x
N
y N y N
y
k w FT
*
2 2 2 1
2
( ) | |
(
|
.
|
\
|
+ |
.
|
\
|
=
= Y X Y X k w FT
C C r s W
N
D C C r s W
N
D Y y
2 2 2 2 2
2
2 2 1 1 1 1 1
2
1 1 2
1 1
B
( ) | | { }
+ + |
.
|
\
|
=
= X Y X Y k w FT
C C r s C r C s W
N
D Y y
1 1 1 1 1
2
1
2
1
2
1
2
1
2
1 1
2
2
2
1
M
{ }
(
(
+ + |
.
|
\
|
+
X Y X Y
C C r s C r C s W
N
D
2 2 2 2 2
2
2
2
2
2
2
2
2
2
2 2
2
1
Case 7
When k = 3
A = 2; B = 2; C = 0;
'
1
= -f /(1-f)
-1
;
'
2
= 0; T1= s1; T2= r1 f (1- f )
-1
;
S1= s2; S2= r2 f (1- f )
-1
;
( ) | |
( )
|
|
.
|
\
|
|
.
|
\
| +
=
=
X f
x X
N
y N y N
y
k w FT
) 1 (
*
2 2 1 1
3
( ) | | ( )
|
.
|
\
|
=
= X Y k w FT
C C s r W
N
D f f Y y
1 1 1 1 1
2
1 1
1
3
1
1 B
(
(
|
.
|
\
|
+
X Y
C C s r W
N
D
2 2 2 2 2
2
2 2
1
( ) | | ( ) {
+ |
.
|
\
|
=
=
2
1
2
1
2 2 2
1
2
1
2
1 1
2
3
1
1
M
X Y k w FT
C r f f C s W
N
D Y y ( ) }
X Y
C C s f f
1 1 1 1
1
1 2
+ {
2
2
2
2
2
2 2
1
Y
C s W
N
D |
.
|
\
|
( ) ( ) }|
X Y X
C C r s f f C r f f
2 2 2 2 2
1 2
2
2
2
2 2
1 2 1
+
Case 8
When k = 4, then
A = 6 ; B= 0 ; C = 0;
'
1
= 0;
'
2
= 0; T
1
= s
1;
T
2
= 0; S
1
= s
2;
S
2
= 0;
( ) | | |
.
|
\
| +
=
=
N
y N y N
y
k w FT
2 2 1 1
4
( ) | | 0 B
4
=
= k w FT
y
( ) | |
(
(
|
.
|
\
|
+
|
.
|
\
|
=
=
2
2
2
2
2
2 2
2
1
2
1
2
1 1
2
4
1 1
V
Y Y
k w FT
C s W
N
D C s W
N
D Y y
NUMERICAL ILLUSTRATION
Consider two populations I and II given in Appendix A
and B. Both populations are divided into two parts as R-
group and NR-group having size N
1
and N
2
respectively
(N = N
1
+ N
2
). The population parameters are displayed in
Tables 1 to 6.
DISCUSSION
Using the imputation for
2
x by the mixture of three
86 Afr. J. Math. Comput. Sci. Res.
Table 1. Parameters of population - I (from Appendix A).
Parameter Entire population For R-group For NR-group
Size N = 180 N1 =100 N2= 80
Mean Y
03 . 159 = Y 60 . 173
1
= Y
81 . 140
2
= Y
Mean X 22 . 113 = X 45 . 128
1
= X
19 . 94
2
= X
Mean square Y 18 . 2205
2
=
Y
S 36 . 2532
2
1
=
Y
S
90 . 1219
2
2
=
Y
S
Mean square X 61 . 1972
2
=
X
S 86 . 2300
2
1
=
X
S
17 924
2
2
. S
X
=
Coefficient of variation of Y 295 . 0 =
Y
C 2899 . 0
1
=
Y
C
248 . 0
2
=
Y
C
Coefficient of variation of X 392 . 0 =
X
C 373 . 0
1
=
X
C
323 . 0
2
=
X
C
Correlation coefficient 897 . 0 =
XY
857 . 0
1
=
XY
956 . 0
2
=
XY
Table 2. Parameters of population - ii (from Appendix B).
Parameter Entire population For R-group For NR-group
Size N = 150 N1 = 90 N2= 60
Mean Y 77 . 63 = Y 33 . 66
1
= Y
92 . 59
2
= Y
Mean X 2 . 29 = X 72 . 30
1
= X
92 . 26
2
= X
Mean Square Y 87 . 299
2
=
Y
S
33 . 349
2
1
=
Y
S
35 . 206
2
2
=
Y
S
Mean Square X 43 . 110
2
=
X
S 67 . 112
2
1
=
X
S
08 100
2
2
. S
X
=
Coefficient of Variation of Y 272 . 0 =
Y
C 282 . 0
1
=
Y
C
2397 . 0
2
=
Y
C
Coefficient of Variation of X 3599 . 0 =
X
C 345 . 0
1
=
X
C 3716 . 0
2
=
X
C
Correlation coefficient 8093 . 0 =
XY
8051 . 0
1
=
XY
8084 . 0
2
=
XY
Let samples of size n = 40 and n = 30 are drawn from population I and II respectively by SRSWOR and post-
stratified into R and NR-groups. The sample values are in Tables 3 and 4.
Table 3. Sample values for population i.
Parameter Entire sample R-group NR-group
Size n = 40 n1 = 28 n2= 12
Fraction f = 0.22 - -
Table 4. Sample values for population ii
Parameter Entire sample R-group NR-group
Size n = 30 n1 = 20 n2= 10
Fraction f = 0.2 - -
population means is performed. The proposed class is in
Equation (3) with bias in theorem 2 and mean squared
error in theorem 3. The class contains some special
imputed estimators for value k = 1, k = 2, k = 3 and k = 4.
A non-imputed class of estimator is also developed which
has bias and mean squared error derived in theorem 4
and 5. This class also has some special cases. The
computation over two population is made whose
description is given in appendix. The two random sample
of size n = 40 and n = 30 are drawn from populations I
and II respectively and post-stratified into two groups.
The Tables 5 and 6 are presenting a numerical
comparison between imputation and non-imputation class
over the two populations in terms of their bias and m. s.
e. The imputation technique (1) is effective because there
is not much increase in the mean squared error due to
Shukla et al. 87
Table 5. Bias and M.S.E. comparisons of ( ) | |
k D FT
y .
Estimator
Population I Population II
Bias M.S.E. Bias M.S.E.
[( y
FT
)
D
]
k=1
-2.8775 18.6000 -1.1874 9.8618
[( y
FT
)
D
]
k=2
3.3127 232.4162 3.4360 41.4794
[( y
FT
)
D
]
k=3
-4.0370 8.4710 -0.6377 6.3540
[( y
FT
)
D
]
k=4
0 43.6500 0 9.2675
Table 6. Bias and MSE comparison of ( ) | |
k w FT
y .
Estimator
Population I Population II
Bias M.S.E. Bias M.S.E.
( ) | |
1 = k w FT
y 0.1433 12.9589 0.1095 6.0552
( ) | |
2 = k w FT
y 0.3141 216.3024 0.1599 46.838
( ) | |
3 = k w FT
y 0.096 4.327 0.031 5.2423
( ) | |
4 = k w FT
y 0 43.65 0 9.2662
imputation. The k = 3 seems to be a good choice. The k =
2 performs worst for both sots of data.
CONCLUSIONS
The technique of mixture of X , 1 X , 2 X performs well
and the imputed factor-type estimator is very close to
non-imputed in terms of mean squared error when k = 1,
2, and 3 holds. The choice k = 3 is better over the other
two. Performance over population II is superior than I. it
seems that factor type lass is able to replace the non-
responded observation in a nice manner.
ACKNOWLEDGEMENT
Authors are thankful to the referee for his valuable
suggestions.
REFERENCES
Agrawal MC, Panda KB (1993). An efficient estimator in post-
stratification. Metron, 51(3-4): 179-187.
Ahmed MS, Al-titi O, Al-Rawi Z, Abu-dayyeh W (2006). Estimation of a
population mean using different imputation method. Stat. Transit.,
7(6): 1247-1264.
Cochran WG (2005). Sampling Techniques. Third Edition, John Willey &
Sons, New Delhi.
Grover RM, Couper MP (1998). Non-response in household surveys.
John Wiley and Sons, New York.
Hansen MH, Hurwitz WN (1946). The problem of non-response in
sample surveys. J. Am. Stat. Assoc., 41: 517-529.
Hansen MH, Hurwitz WN, Madow WG (1993). Sample Survey Methods
and Theory. John Wiley and Sons, New York.
Hinde RL, Chambers RL (1990). Non-response imputation with multiple
source of non-response. J. Off. Stat., 7: 169-179.
Holt D, Smith TMF (1979). Post-stratification. J. R. Stat. Soc., Ser. A.,
143: 33-46.
Jackway PT, Boyce RA (1987). Response including techniques of mail
surveys. Aus. J. Stat., 29: 255-263.
Jagers P (1986). Post-stratification against bias in sampling. Int. Stat.
Rev., 54: 159-167.
Jagers P, Oden A, Trulsoon L (1985). Post-stratification and ratio
estimation. Int. Stat. Rev., 53: 221-238.
Khare BB (1987). Allocation in stratified sampling in presence of non-
response. Metron, 45(I/II): 213-221.
Khot PS (1994). A note on handling non-response in sample surveys. J.
Am. Stat. Assoc., 89: 693-696.
Lessler JT, Kalsbeek WD (1992). Non-response error in surveys. John
Wiley and Sons, New York.
Murthy MN (1976). Sampling Theory and Methods. Statistical
Publishing Society, Calcutta.
Rao JNK, Sitter RR (1995). Variance estimation under two phase
sampling with application to imputation for missing data. Biometrika,
82: 453-460.
Rubin DB (1976). Inference and missing data. Biometrika, 63: 581-593.
Sahoo LN (1984). A note on estimation of the population mean using
two auxiliary variables. Ali. J. Stat., pp. 3-4, 63-66.
Sahoo LN (1986). On a class of unbiased estimators using multi
auxiliary information. J. Ind. Soc. Ag. Stat, 38(3): 379-382.
Sahoo LN, Sahoo RK (2001). Predictive estimation of finite population
mean in two-phase sampling using two auxiliary variables. Ind. Soc.
Ag. Stat., 54(4): 258-264.
88 Afr. J. Math. Comput. Sci. Res.
Sahoo LN, Sahoo J, Mohanty S (1995). New predictive ratio estimator.
J. Ind. Soc. Ag. Stat., 47(3): 240-242.
Shrivastava SK, Jhajj HS (1980). A class of estimators using auxiliary
information for estimation finite population variance. Sankhya C, 42:
87-96.
Shrivastava SK, Jhajj HS (1981). A class of the population mean in a
survey sampling using auxiliary information. Biometrika, 68: 341-343.
Shukla D (2002). F-T estimator under two phase sampling. Metron,
59(1-2): 253-263.
Shukla D, Dubey J (2001). Estimation in mail surveys under PSNR
sampling scheme. Ind. Soc. Ag. Stat., 54(3): 288-302.
Shukla D, Dubey J (2004). On estimation under post-stratified two
phase non-response sampling scheme in mail surveys. Int. J.
Manage. Syst., 20(3): 269-278.
Shukla D, Dubey J (2006). On earmarked strata in post-stratification.
Stat. Transit., 7(5): 1067-1085.
Shukla D, Trivedi M (1999). A new estimator for post-stratified sampling
scheme. Proceedings of NSBA-TA (Bayesian Analysis), Eds. Rajesh
Singh, pp. 206-218.
Shukla D, Trivedi M (2001). Mean estimation in deeply stratified
population under post-stratification. Ind. Soc. Ag. Stat., 54(2): 221-
235.
Shukla D, Trivedi M (2006). Examining stability of variability ratio with
application in post-stratification. Int. J. Manage. Syst., 22(1): 59-70.
Shukla D, Bankey A, Trivedi M (2002). Estimation in post-stratification
using prior information and grouping strategy. Ind. Soc. Ag. Stat.,
55(2): 158-173.
Shukla D, Singh VK, Singh GN (1991). On the use of transformation in
factor type estimator. Metron, 49(1-4): 359-361.
Shukla D, Trivedi M, Singh GN (2006). Post-stratification two-way
deeply stratified population. Stat. Transit., 7(6): 1295-1310.
Singh GN, Singh VK (2001). On the use of auxiliary information in
successive sampling. J. Ind. Soc. Ag. Stat., 54(1): 1-12.
Singh D, Choudhary FS (1986). Theory and Analysis of Sample Survey
Design. Wiley Eastern Limited, New Delhi.
Singh HP (1986). A generalized class of estimators of ratio product and
mean using supplementary information on an auxiliary character in
PPSWR scheme. Guj. Stat. Rev., 13(2): 1-30.
Singh HP, Upadhyaya LN, Lachan R (1987). Unbiased product
estimators. Guj. Stat. Rev., 14(2): 41-50.
Singh S, Horn S (2000). Compromised imputation in survey sampling.
Metrika, 51: 266-276.
Singh VK, Shukla D (1987). One parameter family of factor type ratio
estimator. Metron, 45(1-2): 273-283.
Singh VK, Shukla D (1993). An efficient one-parameter family of factor
type estimator in sample survey. Metron, 51(1-2): 139-159.
Singh VK, Singh GN (1991). Chain type estimator with two auxiliary
variables under double sampling scheme. Metron, 49: 279-289.
Singh VK, Singh GN, Shukla D (1994). A class of chain ratio estimator
with two auxiliary variables under double sampling scheme. Sankhya,
Ser. B., 46(2): 209-221.
Smith TMF (1991). Post-stratification. Statisticians, 40: 323-330.
Sukhatme PV, Sukhatme BV, Sukhatme S, Ashok C (1984). Sampling
Theory of Surveys with Applications. Iowa State University Press,
I.S.A.S. Publication, New Delhi.
Wywial J (2001). Stratification of population after sample selection. Stat.
Transit., 5(2): 327-348.
Shukla et al. 89
Appendix A. Population I (N= 180); R-group: (N1=100).
Y: 110 75 85 165 125 110 85 80 150 165 135 120 140 135 145
X: 80 40 55 130 85 50 35 40 110 115 95 60 70 85 115
Y: 200 135 120 165 150 160 165 145 215 150 145 150 150 195 190
X: 150 85 80 100 25 130 135 105 185 110 95 75 70 165 160
Y: 175 160 165 175 185 205 140 105 125 230 230 255 275 145 125
X: 145 110 135 145 155 175 80 75 65 170 170 190 205 105 85
Y: 110 110 120 230 220 280 275 220 145 155 170 195 170 185 195
X: 75 80 90 165 160 205 215 190 105 115 135 145 135 110 145
Y: 180 150 185 165 285 150 235 125 165 135 130 245 255 280 150
X: 135 110 135 115 125 205 100 195 85 115 75 190 205 210 105
Y: 205 180 150 205 220 240 260 185 150 155 115 115 220 215 230
X: 110 105 110 175 180 215 225 110 90 95 85 75 175 185 190
Y: 210 145 135 250 265 275 205 195 180 115
X: 170 85 95 190 215 200 165 155 150 175
NR-group: (N2=80)
Y: 85 75 115 165 140 110 115 13.5 120 125 120 150 145 90 105
X: 55 40 65 115 90 55 60 65 70 75 80 120 105 45 65
Y: 110 90 155 130 120 95 100 125 140 155 160 145 90 90 95
X: 70 60 85 95 80 55 60 75 90 105 125 95 45 55 65
Y: 115 140 180 170 175 190 160 155 175 195 90 90 80 90 80
X: 75 105 120 115 125 135 110 115 135 145 45 55 50 60 50
Y: 105 125 110 120 130 145 160 170 180 `145 130 195 200 160 110
X: 65 75 70 80 85 105 110 115 130 95 65 135 130 115 55
Y: 155 190 150 180 200 160 155 170 195 200 150 165 155 180 200
X: 115 130 110 120 125 145 120 105 100 95 90 105 125 130 145
Y: 160 155 170 195 200
X: 120 115 120 135 150
Appendix B. Population II (N=150); R-group (N1=90).
Y: 90 75 70 85 95 55 65 80 65 50 45 55 60 60 95
X: 30 35 30 40 45 25 40 50 35 30 15 20 25 30 40
Y: 100 40 45 55 35 45 35 55 85 95 65 75 70 80 65
X: 50 10 25 25 10 15 10 25 35 55 35 40 30 45 40
Y: 90 95 80 85 55 60 75 85 80 65 35 40 95 100 55
X: 40 50 35 45 35 25 30 40 25 35 10 15 45 45 25
Y: 45 40 40 35 55 75 80 80 85 55 45 70 80 90 55
X: 15 15 20 10 30 25 30 40 35 20 25 30 40 45 30
Y: 65 60 75 75 85 95 90 90 45 40 45 55 60 65 60
X: 25 40 35 30 40 35 40 35 15 25 15 30 30 25 20
Y: 75 70 40 55 75 45 55 60 85 55 60 70 75 65 80
X: 25 20 35 30 45 10 30 25 40 15 25 30 35 30 45
NR-group (N2=60)
Y: 40 90 95 70 60 65 85 55 45 60 65 60 55 55 45
X: 10 30 30 30 25 30 40 25 15 20 30 30 35 25 20
Y: 65 80 55 65 75 55 50 55 60 45 40 75 75 45 70
X: 35 45 30 30 40 15 15 20 30 15 10 40 45 10 30
Y: 65 70 55 35 35 50 55 35 55 60 30 35 45 55 65
X: 30 40 30 10 15 25 30 15 20 30 10 20 15 30 30
Y: 75 65 70 65 70 45 55 60 85 55 60 70 75 65 80
X: 30 35 40 25 45 10 30 25 40 15 25 30 35 30 45