American Time Use Survey (ATUS): CATI Paradata, 2010
- Belli, Robert F.
- Cordova Cazar, Ana Lucia
- Eck, Adam
- Olson, Kristen
- ATUS Paradata, 2010 (Alternative Title)
- Archival Version (Subtitle)
- American Time Use Survey (ATUS) Series
- Inter-University Consortium for Political and Social Research
- National Science Foundation
data complexity; interviewer observations; paradata; time utilization
AbstractThe annual American Time Use Survey (ATUS) collects information on people in the United States and what activities they spend their time in during a 24 hour period. Questions are not only asked about what activities are engaged in, but also with who is with them at that time, for how long do they spend doing the activity, and where the activity takes place. The current study contains the observed paradata derived from the original Blaise audit trails of the full 2010 ATUS public-use data. The dataset contains data describing the interaction by interviewers with the CATI instrument while entering responses provided by respondents. The file contains 18 variables and 2,061,889 cases.
AbstractThe overall aim of this study was to improve data quality assessment through the use of paradata and substantive answers from the 2010 American Time Use Survey (ATUS). Specifically, the relationship between data quality and interviewing complexity is explored. The specific objectives of this research were: (a) To investigate the data quality of the 2010 ATUS through the construction of data quality index;; (b) To explore whether paradata can serve as a tool to assess interview complexity;; (c) To examine whether interview complexity predicts data quality in the ATUS.; In the study data quality and interview complexity were considered latent constructs that were measured by observed variables that include both survey responses and paradata. Data quality was measured through observed indicators including the total number of activities reported, the total time spent on responding to the interview, 'rounding' instances, item non-response instances (memory gaps), inadequate responses (i.e., instances in which the respondent reported an activity with insufficient detail such that the activity could not be coded), failures in reporting a basic daily activity (e.g., eating or sleeping), and failures in providing additional details (where and with whom the activity took place). Interview complexity was exclusively measured through paradata variables, including the total number of entries per respondent, the proportion of secondary activities reported, the number of times reported activities and their locations where edited, and the total number of verbatim reports.
MethodsParadata was derived from the original Blaise audit trails during data collection of the 2010 survey. The paradata describes the interaction between the survey interviewer and the CATI instrument during administration of the survey. These "audit trails" are non-rectangular text data capturing key strokes and navigation during the course of the survey interview.
MethodsThe data file contains 18 variables. Key variables in the file include the time each activity started (TIMEBEGAN), ended (TIMEENDED), and lasted (TIMEELAPSED) measured in seconds. The start and end times are very precise (M:D:Y / H:M:S).
MethodsICPSR data undergo a confidentiality review and are altered when necessary to limit the risk of disclosure. ICPSR also routinely creates ready-to-go data files along with setups in the major statistical software formats as well as standard codebooks to accompany the data. In addition to these procedures, ICPSR performed the following processing steps for this data collection: Checked for undocumented or out-of-range codes..
MethodsResponse Rates: 56.9 percent
- DS1: Dataset
Time period: 2010
Collection date: 2010-01-01--2010-12-31
2010-01-01 / 2010-12-31
computer-assisted telephone interview (CATI)
- 37318 (Type: ICPSR Study Number)
Is previous version of
Update Metadata: 2019-06-17 | Issue Number: 1 | Registration Date: 2019-06-17