U.S. Census Bureau

General Methods and Algorithms for Modeling and Imputing Discrete Data under a Variety of Constraints

William E. Winkler

KEY WORDS: Data Quality, Loglinear Model Fit, Missing Data, Convex Constraints

ABSTRACT

Loglinear modeling methods have become quite straightforward to apply to discrete data X. The models for missing data involve minor extensions of hot-deck methods (Little and Rubin 2002). Edits are structural zeros that forbid certain patterns. Winkler (2003) provided the theory for connecting edit with imputation. In this paper, we give methods and algorithms for modeling/edit/imputation under linear and convex constraints. The methods can be used for statistical matching (D’Orazio, Di Zio, and Scanu 2006), edit/imputation in which models are also controlled for external constraints such as in benchmark data, and for creating synthetic data with significantly reduced risk of re-identification (Winkler 2007).

CITATION: Winkler, W.E. General Methods and Algorithms for Modeling and Imputing Discrete Data under a Variety of Constraints. Statistical Research Division Research Report Series (Statistics #2008-08). U.S. Census Bureau.

Source: U.S. Census Bureau, Statistical Research Division

Created: October 3, 2008
Last revised: October 3, 2008