635 lines
18 KiB
Python
635 lines
18 KiB
Python
"""
|
|
Conditional logistic, Poisson, and multinomial logit regression
|
|
"""
|
|
|
|
import numpy as np
|
|
import statsmodels.base.model as base
|
|
import statsmodels.regression.linear_model as lm
|
|
import statsmodels.base.wrapper as wrap
|
|
from statsmodels.discrete.discrete_model import (MultinomialResults,
|
|
MultinomialResultsWrapper)
|
|
import collections
|
|
import warnings
|
|
import itertools
|
|
|
|
|
|
class _ConditionalModel(base.LikelihoodModel):
|
|
|
|
def __init__(self, endog, exog, missing='none', **kwargs):
|
|
|
|
if "groups" not in kwargs:
|
|
raise ValueError("'groups' is a required argument")
|
|
groups = kwargs["groups"]
|
|
|
|
if groups.size != endog.size:
|
|
msg = "'endog' and 'groups' should have the same dimensions"
|
|
raise ValueError(msg)
|
|
|
|
if exog.shape[0] != endog.size:
|
|
msg = "The leading dimension of 'exog' should equal the length of 'endog'"
|
|
raise ValueError(msg)
|
|
|
|
super().__init__(
|
|
endog, exog, missing=missing, **kwargs)
|
|
|
|
if self.data.const_idx is not None:
|
|
msg = ("Conditional models should not have an intercept in the " +
|
|
"design matrix")
|
|
raise ValueError(msg)
|
|
|
|
exog = self.exog
|
|
self.k_params = exog.shape[1]
|
|
|
|
# Get the row indices for each group
|
|
row_ix = {}
|
|
for i, g in enumerate(groups):
|
|
if g not in row_ix:
|
|
row_ix[g] = []
|
|
row_ix[g].append(i)
|
|
|
|
# Split the data into groups and remove groups with no variation
|
|
endog, exog = np.asarray(endog), np.asarray(exog)
|
|
offset = kwargs.get("offset")
|
|
self._endog_grp = []
|
|
self._exog_grp = []
|
|
self._groupsize = []
|
|
if offset is not None:
|
|
offset = np.asarray(offset)
|
|
self._offset_grp = []
|
|
self._offset = []
|
|
self._sumy = []
|
|
self.nobs = 0
|
|
drops = [0, 0]
|
|
for g, ix in row_ix.items():
|
|
y = endog[ix].flat
|
|
if np.std(y) == 0:
|
|
drops[0] += 1
|
|
drops[1] += len(y)
|
|
continue
|
|
self.nobs += len(y)
|
|
self._endog_grp.append(y)
|
|
if offset is not None:
|
|
self._offset_grp.append(offset[ix])
|
|
self._groupsize.append(len(y))
|
|
self._exog_grp.append(exog[ix, :])
|
|
self._sumy.append(np.sum(y))
|
|
|
|
if drops[0] > 0:
|
|
msg = ("Dropped %d groups and %d observations for having " +
|
|
"no within-group variance") % tuple(drops)
|
|
warnings.warn(msg)
|
|
|
|
# This can be pre-computed
|
|
if offset is not None:
|
|
self._endofs = []
|
|
for k, ofs in enumerate(self._offset_grp):
|
|
self._endofs.append(np.dot(self._endog_grp[k], ofs))
|
|
|
|
# Number of groups
|
|
self._n_groups = len(self._endog_grp)
|
|
|
|
# These are the sufficient statistics
|
|
self._xy = []
|
|
self._n1 = []
|
|
for g in range(self._n_groups):
|
|
self._xy.append(np.dot(self._endog_grp[g], self._exog_grp[g]))
|
|
self._n1.append(np.sum(self._endog_grp[g]))
|
|
|
|
def hessian(self, params):
|
|
|
|
from statsmodels.tools.numdiff import approx_fprime
|
|
hess = approx_fprime(params, self.score)
|
|
hess = np.atleast_2d(hess)
|
|
return hess
|
|
|
|
def fit(self,
|
|
start_params=None,
|
|
method='BFGS',
|
|
maxiter=100,
|
|
full_output=True,
|
|
disp=False,
|
|
fargs=(),
|
|
callback=None,
|
|
retall=False,
|
|
skip_hessian=False,
|
|
**kwargs):
|
|
|
|
rslt = super().fit(
|
|
start_params=start_params,
|
|
method=method,
|
|
maxiter=maxiter,
|
|
full_output=full_output,
|
|
disp=disp,
|
|
skip_hessian=skip_hessian)
|
|
|
|
crslt = ConditionalResults(self, rslt.params, rslt.cov_params(), 1)
|
|
crslt.method = method
|
|
crslt.nobs = self.nobs
|
|
crslt.n_groups = self._n_groups
|
|
crslt._group_stats = [
|
|
"%d" % min(self._groupsize),
|
|
"%d" % max(self._groupsize),
|
|
"%.1f" % np.mean(self._groupsize)
|
|
]
|
|
rslt = ConditionalResultsWrapper(crslt)
|
|
return rslt
|
|
|
|
def fit_regularized(self,
|
|
method="elastic_net",
|
|
alpha=0.,
|
|
start_params=None,
|
|
refit=False,
|
|
**kwargs):
|
|
"""
|
|
Return a regularized fit to a linear regression model.
|
|
|
|
Parameters
|
|
----------
|
|
method : {'elastic_net'}
|
|
Only the `elastic_net` approach is currently implemented.
|
|
alpha : scalar or array_like
|
|
The penalty weight. If a scalar, the same penalty weight
|
|
applies to all variables in the model. If a vector, it
|
|
must have the same length as `params`, and contains a
|
|
penalty weight for each coefficient.
|
|
start_params : array_like
|
|
Starting values for `params`.
|
|
refit : bool
|
|
If True, the model is refit using only the variables that
|
|
have non-zero coefficients in the regularized fit. The
|
|
refitted model is not regularized.
|
|
**kwargs
|
|
Additional keyword argument that are used when fitting the model.
|
|
|
|
Returns
|
|
-------
|
|
Results
|
|
A results instance.
|
|
"""
|
|
|
|
from statsmodels.base.elastic_net import fit_elasticnet
|
|
|
|
if method != "elastic_net":
|
|
raise ValueError("method for fit_regularized must be elastic_net")
|
|
|
|
defaults = {"maxiter": 50, "L1_wt": 1, "cnvrg_tol": 1e-10,
|
|
"zero_tol": 1e-10}
|
|
defaults.update(kwargs)
|
|
|
|
return fit_elasticnet(self, method=method,
|
|
alpha=alpha,
|
|
start_params=start_params,
|
|
refit=refit,
|
|
**defaults)
|
|
|
|
# Override to allow groups to be passed as a variable name.
|
|
@classmethod
|
|
def from_formula(cls,
|
|
formula,
|
|
data,
|
|
subset=None,
|
|
drop_cols=None,
|
|
*args,
|
|
**kwargs):
|
|
|
|
try:
|
|
groups = kwargs["groups"]
|
|
del kwargs["groups"]
|
|
except KeyError:
|
|
raise ValueError("'groups' is a required argument")
|
|
|
|
if isinstance(groups, str):
|
|
groups = data[groups]
|
|
|
|
if "0+" not in formula.replace(" ", ""):
|
|
warnings.warn("Conditional models should not include an intercept")
|
|
|
|
model = super().from_formula(
|
|
formula, data=data, groups=groups, *args, **kwargs)
|
|
|
|
return model
|
|
|
|
|
|
class ConditionalLogit(_ConditionalModel):
|
|
"""
|
|
Fit a conditional logistic regression model to grouped data.
|
|
|
|
Every group is implicitly given an intercept, but the model is fit using
|
|
a conditional likelihood in which the intercepts are not present. Thus,
|
|
intercept estimates are not given, but the other parameter estimates can
|
|
be interpreted as being adjusted for any group-level confounders.
|
|
|
|
Parameters
|
|
----------
|
|
endog : array_like
|
|
The response variable, must contain only 0 and 1.
|
|
exog : array_like
|
|
The array of covariates. Do not include an intercept
|
|
in this array.
|
|
groups : array_like
|
|
Codes defining the groups. This is a required keyword parameter.
|
|
"""
|
|
|
|
def __init__(self, endog, exog, missing='none', **kwargs):
|
|
|
|
super().__init__(
|
|
endog, exog, missing=missing, **kwargs)
|
|
|
|
if np.any(np.unique(self.endog) != np.r_[0, 1]):
|
|
msg = "endog must be coded as 0, 1"
|
|
raise ValueError(msg)
|
|
|
|
self.K = self.exog.shape[1]
|
|
# i.e. self.k_params, for compatibility with MNLogit
|
|
|
|
def loglike(self, params):
|
|
|
|
ll = 0
|
|
for g in range(len(self._endog_grp)):
|
|
ll += self.loglike_grp(g, params)
|
|
|
|
return ll
|
|
|
|
def score(self, params):
|
|
|
|
score = 0
|
|
for g in range(self._n_groups):
|
|
score += self.score_grp(g, params)
|
|
|
|
return score
|
|
|
|
def _denom(self, grp, params, ofs=None):
|
|
|
|
if ofs is None:
|
|
ofs = 0
|
|
|
|
exb = np.exp(np.dot(self._exog_grp[grp], params) + ofs)
|
|
|
|
# In the recursions, f may be called multiple times with the
|
|
# same arguments, so we memoize the results.
|
|
memo = {}
|
|
|
|
def f(t, k):
|
|
if t < k:
|
|
return 0
|
|
if k == 0:
|
|
return 1
|
|
|
|
try:
|
|
return memo[(t, k)]
|
|
except KeyError:
|
|
pass
|
|
|
|
v = f(t - 1, k) + f(t - 1, k - 1) * exb[t - 1]
|
|
memo[(t, k)] = v
|
|
|
|
return v
|
|
|
|
return f(self._groupsize[grp], self._n1[grp])
|
|
|
|
def _denom_grad(self, grp, params, ofs=None):
|
|
|
|
if ofs is None:
|
|
ofs = 0
|
|
|
|
ex = self._exog_grp[grp]
|
|
exb = np.exp(np.dot(ex, params) + ofs)
|
|
|
|
# s may be called multiple times in the recursions with the
|
|
# same arguments, so memoize the results.
|
|
memo = {}
|
|
|
|
def s(t, k):
|
|
|
|
if t < k:
|
|
return 0, np.zeros(self.k_params)
|
|
if k == 0:
|
|
return 1, 0
|
|
|
|
try:
|
|
return memo[(t, k)]
|
|
except KeyError:
|
|
pass
|
|
|
|
h = exb[t - 1]
|
|
a, b = s(t - 1, k)
|
|
c, e = s(t - 1, k - 1)
|
|
d = c * h * ex[t - 1, :]
|
|
|
|
u, v = a + c * h, b + d + e * h
|
|
memo[(t, k)] = (u, v)
|
|
|
|
return u, v
|
|
|
|
return s(self._groupsize[grp], self._n1[grp])
|
|
|
|
def loglike_grp(self, grp, params):
|
|
|
|
ofs = None
|
|
if hasattr(self, 'offset'):
|
|
ofs = self._offset_grp[grp]
|
|
|
|
llg = np.dot(self._xy[grp], params)
|
|
|
|
if ofs is not None:
|
|
llg += self._endofs[grp]
|
|
|
|
llg -= np.log(self._denom(grp, params, ofs))
|
|
|
|
return llg
|
|
|
|
def score_grp(self, grp, params):
|
|
|
|
ofs = 0
|
|
if hasattr(self, 'offset'):
|
|
ofs = self._offset_grp[grp]
|
|
|
|
d, h = self._denom_grad(grp, params, ofs)
|
|
return self._xy[grp] - h / d
|
|
|
|
|
|
class ConditionalPoisson(_ConditionalModel):
|
|
"""
|
|
Fit a conditional Poisson regression model to grouped data.
|
|
|
|
Every group is implicitly given an intercept, but the model is fit using
|
|
a conditional likelihood in which the intercepts are not present. Thus,
|
|
intercept estimates are not given, but the other parameter estimates can
|
|
be interpreted as being adjusted for any group-level confounders.
|
|
|
|
Parameters
|
|
----------
|
|
endog : array_like
|
|
The response variable
|
|
exog : array_like
|
|
The covariates
|
|
groups : array_like
|
|
Codes defining the groups. This is a required keyword parameter.
|
|
"""
|
|
|
|
def loglike(self, params):
|
|
|
|
ofs = None
|
|
if hasattr(self, 'offset'):
|
|
ofs = self._offset_grp
|
|
|
|
ll = 0.0
|
|
|
|
for i in range(len(self._endog_grp)):
|
|
|
|
xb = np.dot(self._exog_grp[i], params)
|
|
if ofs is not None:
|
|
xb += ofs[i]
|
|
exb = np.exp(xb)
|
|
y = self._endog_grp[i]
|
|
ll += np.dot(y, xb)
|
|
s = exb.sum()
|
|
ll -= self._sumy[i] * np.log(s)
|
|
|
|
return ll
|
|
|
|
def score(self, params):
|
|
|
|
ofs = None
|
|
if hasattr(self, 'offset'):
|
|
ofs = self._offset_grp
|
|
|
|
score = 0.0
|
|
|
|
for i in range(len(self._endog_grp)):
|
|
|
|
x = self._exog_grp[i]
|
|
xb = np.dot(x, params)
|
|
if ofs is not None:
|
|
xb += ofs[i]
|
|
exb = np.exp(xb)
|
|
s = exb.sum()
|
|
y = self._endog_grp[i]
|
|
score += np.dot(y, x)
|
|
score -= self._sumy[i] * np.dot(exb, x) / s
|
|
|
|
return score
|
|
|
|
|
|
class ConditionalResults(base.LikelihoodModelResults):
|
|
def __init__(self, model, params, normalized_cov_params, scale):
|
|
|
|
super().__init__(
|
|
model,
|
|
params,
|
|
normalized_cov_params=normalized_cov_params,
|
|
scale=scale)
|
|
|
|
def summary(self, yname=None, xname=None, title=None, alpha=.05):
|
|
"""
|
|
Summarize the fitted model.
|
|
|
|
Parameters
|
|
----------
|
|
yname : str, optional
|
|
Default is `y`
|
|
xname : list[str], optional
|
|
Names for the exogenous variables, default is "var_xx".
|
|
Must match the number of parameters in the model
|
|
title : str, optional
|
|
Title for the top table. If not None, then this replaces the
|
|
default title
|
|
alpha : float
|
|
Significance level for the confidence intervals
|
|
|
|
Returns
|
|
-------
|
|
smry : Summary instance
|
|
This holds the summary tables and text, which can be printed or
|
|
converted to various output formats.
|
|
|
|
See Also
|
|
--------
|
|
statsmodels.iolib.summary.Summary : class to hold summary
|
|
results
|
|
"""
|
|
|
|
top_left = [
|
|
('Dep. Variable:', None),
|
|
('Model:', None),
|
|
('Log-Likelihood:', None),
|
|
('Method:', [self.method]),
|
|
('Date:', None),
|
|
('Time:', None),
|
|
]
|
|
|
|
top_right = [
|
|
('No. Observations:', None),
|
|
('No. groups:', [self.n_groups]),
|
|
('Min group size:', [self._group_stats[0]]),
|
|
('Max group size:', [self._group_stats[1]]),
|
|
('Mean group size:', [self._group_stats[2]]),
|
|
]
|
|
|
|
if title is None:
|
|
title = "Conditional Logit Model Regression Results"
|
|
|
|
# create summary tables
|
|
from statsmodels.iolib.summary import Summary
|
|
smry = Summary()
|
|
smry.add_table_2cols(
|
|
self,
|
|
gleft=top_left,
|
|
gright=top_right, # [],
|
|
yname=yname,
|
|
xname=xname,
|
|
title=title)
|
|
smry.add_table_params(
|
|
self, yname=yname, xname=xname, alpha=alpha, use_t=self.use_t)
|
|
|
|
return smry
|
|
|
|
class ConditionalMNLogit(_ConditionalModel):
|
|
"""
|
|
Fit a conditional multinomial logit model to grouped data.
|
|
|
|
Parameters
|
|
----------
|
|
endog : array_like
|
|
The dependent variable, must be integer-valued, coded
|
|
0, 1, ..., c-1, where c is the number of response
|
|
categories.
|
|
exog : array_like
|
|
The independent variables.
|
|
groups : array_like
|
|
Codes defining the groups. This is a required keyword parameter.
|
|
|
|
Notes
|
|
-----
|
|
Equivalent to femlogit in Stata.
|
|
|
|
References
|
|
----------
|
|
Gary Chamberlain (1980). Analysis of covariance with qualitative
|
|
data. The Review of Economic Studies. Vol. 47, No. 1, pp. 225-238.
|
|
"""
|
|
|
|
def __init__(self, endog, exog, missing='none', **kwargs):
|
|
|
|
super().__init__(
|
|
endog, exog, missing=missing, **kwargs)
|
|
|
|
# endog must be integers
|
|
self.endog = self.endog.astype(int)
|
|
|
|
self.k_cat = self.endog.max() + 1
|
|
self.df_model = (self.k_cat - 1) * self.exog.shape[1]
|
|
self.df_resid = self.nobs - self.df_model
|
|
self._ynames_map = {j: str(j) for j in range(self.k_cat)}
|
|
self.J = self.k_cat # Unfortunate name, needed for results
|
|
self.K = self.exog.shape[1] # for compatibility with MNLogit
|
|
|
|
if self.endog.min() < 0:
|
|
msg = "endog may not contain negative values"
|
|
raise ValueError(msg)
|
|
|
|
grx = collections.defaultdict(list)
|
|
for k, v in enumerate(self.groups):
|
|
grx[v].append(k)
|
|
self._group_labels = list(grx.keys())
|
|
self._group_labels.sort()
|
|
self._grp_ix = [grx[k] for k in self._group_labels]
|
|
|
|
def fit(self,
|
|
start_params=None,
|
|
method='BFGS',
|
|
maxiter=100,
|
|
full_output=True,
|
|
disp=False,
|
|
fargs=(),
|
|
callback=None,
|
|
retall=False,
|
|
skip_hessian=False,
|
|
**kwargs):
|
|
|
|
if start_params is None:
|
|
q = self.exog.shape[1]
|
|
c = self.k_cat - 1
|
|
start_params = np.random.normal(size=q * c)
|
|
|
|
# Do not call super(...).fit because it cannot handle the 2d-params.
|
|
rslt = base.LikelihoodModel.fit(
|
|
self,
|
|
start_params=start_params,
|
|
method=method,
|
|
maxiter=maxiter,
|
|
full_output=full_output,
|
|
disp=disp,
|
|
skip_hessian=skip_hessian)
|
|
|
|
rslt.params = rslt.params.reshape((self.exog.shape[1], -1))
|
|
rslt = MultinomialResults(self, rslt)
|
|
|
|
# Not clear what the null likelihood should be, there is no intercept
|
|
# so the null model is not clearly defined. This is needed for summary
|
|
# to work.
|
|
rslt.set_null_options(llnull=np.nan)
|
|
|
|
return MultinomialResultsWrapper(rslt)
|
|
|
|
def loglike(self, params):
|
|
|
|
q = self.exog.shape[1]
|
|
c = self.k_cat - 1
|
|
|
|
pmat = params.reshape((q, c))
|
|
pmat = np.concatenate((np.zeros((q, 1)), pmat), axis=1)
|
|
lpr = np.dot(self.exog, pmat)
|
|
|
|
ll = 0.0
|
|
for ii in self._grp_ix:
|
|
x = lpr[ii, :]
|
|
jj = np.arange(x.shape[0], dtype=int)
|
|
y = self.endog[ii]
|
|
denom = 0.0
|
|
for p in itertools.permutations(y):
|
|
denom += np.exp(x[(jj, p)].sum())
|
|
ll += x[(jj, y)].sum() - np.log(denom)
|
|
|
|
return ll
|
|
|
|
|
|
def score(self, params):
|
|
|
|
q = self.exog.shape[1]
|
|
c = self.k_cat - 1
|
|
|
|
pmat = params.reshape((q, c))
|
|
pmat = np.concatenate((np.zeros((q, 1)), pmat), axis=1)
|
|
lpr = np.dot(self.exog, pmat)
|
|
|
|
grad = np.zeros((q, c))
|
|
for ii in self._grp_ix:
|
|
x = lpr[ii, :]
|
|
jj = np.arange(x.shape[0], dtype=int)
|
|
y = self.endog[ii]
|
|
denom = 0.0
|
|
denomg = np.zeros((q, c))
|
|
for p in itertools.permutations(y):
|
|
v = np.exp(x[(jj, p)].sum())
|
|
denom += v
|
|
for i, r in enumerate(p):
|
|
if r != 0:
|
|
denomg[:, r - 1] += v * self.exog[ii[i], :]
|
|
|
|
for i, r in enumerate(y):
|
|
if r != 0:
|
|
grad[:, r - 1] += self.exog[ii[i], :]
|
|
|
|
grad -= denomg / denom
|
|
|
|
return grad.flatten()
|
|
|
|
|
|
|
|
class ConditionalResultsWrapper(lm.RegressionResultsWrapper):
|
|
pass
|
|
|
|
|
|
wrap.populate_wrapper(ConditionalResultsWrapper, ConditionalResults)
|