Skip to contents

Empirical Likelihood Method

To consider the empirical likelihood under non-ignorable missing data, one has to discuss the distribution function of multidimensional random variable (𝐗,Y)(\boldsymbol{X},Y) which is determined by its distribution function F(x,y)F(x,y). There are no assumptions on F(x,y)F(x,y) (except the fact that it has to fit CDF assumptions) but there is a setting in Kim and Shao (2013), that 𝔼{U(𝛉0;X,Y)}=0\begin{equation}\label{eq: setting on (X,Y)} \mathop{\mathrm{\mathbb{E}}}\{U(\boldsymbol{\theta}_0; X,Y)\} = 0 \end{equation} which is a mm-dimensional, linearly independent vector, U(β‹…)βˆˆπ’ž2U(\cdot) \in \mathcal{C}^2 and 𝛉0βˆˆΞ©βŠ†β„p\boldsymbol{\theta}_0 \in \Omega \subseteq \mathbb{R}^p. if m=pm = p, then a consistent estimator of 𝛉0\boldsymbol{\theta}_0 is a solution of: βˆ‘i=1nU(𝛉,𝐱i,y).\begin{equation}\label{eq: consistent estimator of btheta under m = p} \sum_{i=1}^{n} U(\boldsymbol{\theta}, \boldsymbol{x}_i, y). \end{equation} When m>pm > p the model is called , thus one has to adjust the methodology which results in different optimization problem, since might not provide solution at all. This adjustment is also proposed in Kim and Shao (2013) and the EL approach is concentrated around finding a solution 𝛉\boldsymbol{\theta} that maximizes the empirical likelihood function of 𝛉\boldsymbol{\theta}: L(𝛉)=arg max𝛉{∏i=1npi:pi>0,βˆ‘i=1npiβˆ§βˆ‘i=1npiU(𝛉;𝐱i,yi)=0}.\begin{equation}\label{eq: EL solution for m>p} L(\boldsymbol{\theta}) = \mathop{\mathrm{arg\,max}}_{\boldsymbol{\theta}}\left\{\prod_{i=1}^n p_i: p_i>0, \sum_{i=1}^np_i \wedge \sum_{i=1}^n p_i U(\boldsymbol{\theta}; \boldsymbol{x}_i, y_i) = 0\right\}. \end{equation} Using the Lagrange multiplier method, lets denote pΜ‚i(𝛉)=1n11+λ̂𝛉′U(ΞΈ;𝐱i,yi).\begin{equation} \hat{p}_i(\boldsymbol{\theta}) =\frac{1}{n} \frac{1}{1+\hat{\lambda}_{\boldsymbol{\theta}}'U(\theta; \boldsymbol{x}_i, y_i)}. \end{equation} Thus, the MEL estimator is being obtained by maximizing le(𝛉)=βˆ‘i=1nlog{pΜ‚i(𝛉)}.\begin{equation} \label{eq: empirical likelihood method estimator} l_e(\boldsymbol{\theta}) = \sum_{i=1}^n\log\{\hat{p}_i(\boldsymbol{\theta})\}. \end{equation} When dealing with any missingness in data, (recall, that we do have 𝐱i\boldsymbol{x}_i for any individual and yiy_i is only observed for respondents) the scoring of propensity is applied in such fashion:

  • let RR be a response indicator (Ri=1R_i=1 if yiy_i is observed and Ri=0R_i=0 if yiy_i is unobserved),
  • denote response propensity as Ο€(𝐱;Ο•)=P(R=1|𝐱)\pi(\boldsymbol{x};\phi) = P(R =1|\boldsymbol{x}) (MNAR mechanism provides lack of yiy_i in conditioning),
  • let S(Ο•;R,𝐱)S(\phi; R, \boldsymbol{x}) be the score function of Ο•\phi s.t. 𝔼{S(Ο•;R,𝐱)}=0\mathop{\mathrm{\mathbb{E}}}\{S(\phi; R, \boldsymbol{x})\} = 0 holds.

Thus, 𝔼{S(Ο•;R,𝐱)}=0\mathop{\mathrm{\mathbb{E}}}\{S(\phi; R, \boldsymbol{x})\} = 0 is one of the moment conditions, along with second one of form: 𝔼{RΟ€(𝐗;Ο•)U(𝛉,𝐗,Y)}=0\begin{equation}\label{eq: second moment condition for EL by PS} \mathop{\mathrm{\mathbb{E}}}\left\{\frac{R}{\pi(\boldsymbol{X};\phi)}U(\boldsymbol{\theta},\boldsymbol{X}, Y)\right\} = 0 \end{equation} Those two moment conditions allows us to perform propensity-score-based MELE by maximizing: L(𝛉,Ο•)=arg max{∏i=1npi:𝐩∈B(𝛉,Ο•)},\begin{equation}\label{eq: MELE by PS} L(\boldsymbol{\theta}, \phi) = \mathop{\mathrm{arg\,max}}\left\{\prod_{i=1}^n p_i: \boldsymbol{p}\in B(\boldsymbol{\theta}, \phi) \right\}, \end{equation} where B(𝛉,Ο•)={𝐩:pi>0,βˆ‘i=1npi,βˆ‘i=1npiRiΟ€(𝐱i;Ο•)U(𝛉;𝐱i,yi)=0βˆ‘i=1npiS(Ο•;Ri,𝐱i)=0}.\begin{equation} B(\boldsymbol{\theta}, \phi) = \left\{\boldsymbol{p}: p_i>0, \sum_{i=1}^np_i, \sum_{i=1}^n p_i\frac{R_i}{\pi(\boldsymbol{x}_i;\phi)} U(\boldsymbol{\theta}; \boldsymbol{x}_i, y_i) = 0\, \sum_{i=1}^n p_i S(\phi; R_i, \boldsymbol{x}_i) = 0 \right\}. \end{equation}

References

Kim, J. K., and J. Shao. 2013. Statistical Methods for Handling Incomplete Data. Chapman & Hall Book. Taylor & Francis. https://books.google.pl/books?id=MY4AAAAAQBAJ.