www.fgks.org   »   [go: up one dir, main page]

Skip To Content
Click for DHHS Home Page
Click for the SAMHSA Home Page
Click for the OAS Drug Abuse Statistics Home Page
Click for What's New
Click for Recent Reports and HighlightsClick for Information by Topic Click for OAS Data Systems and more Pubs Click for Data on Specific Drugs of Use Click for Short Reports and Facts Click for Frequently Asked Questions Click for Publications Click to send OAS Comments, Questions and Requests Click for OAS Home Page Click for Substance Abuse and Mental Health Services Administration Home Page Click to Search Our Site


2007 National Survey on Drug Use & Health:  National Results

Go to the Table Of Contents

Appendix A: Description of the Survey

A.1 Sample Design

The 2007 National Survey on Drug Use and Health (NSDUH)3 is part of a coordinated 5-year sample design providing estimates for all 50 States plus the District of Columbia for the years 2005 through 2009. The respondent universe is the civilian, noninstitutionalized population aged 12 years old or older residing within the United States. The survey includes persons living in noninstitutionalized group quarters (e.g., shelters, rooming/boarding houses, college dormitories, migratory workers' camps, halfway houses), and civilians living on military bases. Persons excluded from the survey include persons with no fixed household address (e.g., homeless and/or transient persons not in shelters), active-duty military personnel, and residents of institutional group quarters, such as correctional facilities, nursing homes, mental institutions, and long-term hospitals.

Although there is no planned overlap with the 1999 through 2004 samples, a coordinated design for 2005 through 2009 facilitates 50.0 percent overlap in second-stage units (area segments) within each successive 2-year period from 2005 through 2009. Because the 2005 design enables estimates to be developed by State in all 50 States plus the District of Columbia, States may be viewed as the first level of stratification and as a reporting variable.

For the 50-State design, 8 States were designated as large sample States (California, Florida, Illinois, Michigan, New York, Ohio, Pennsylvania, and Texas) with target sample sizes of 3,600. In 2007, sample sizes in these States ranged from 3,557 to 3,699. For the remaining 42 States and the District of Columbia, the target sample size was 900. Sample sizes in these States ranged from 824 to 974 in 2007. This approach ensures there is sufficient sample in every State to support small area estimation (SAE)4 while at the same time maintaining efficiency for national estimates.

States were first stratified into a total of 900 State sampling (SS) regions (48 regions in each large sample State and 12 regions in each small sample State). These regions were contiguous geographic areas designed to yield the same number of interviews on average.5 Unlike the 1999 through 2001 NHSDAs and the 2002 through 2004 NSDUHs in which the first-stage sampling units were clusters of census blocks called area segments, the first stage of selection for the 2005 through 2009 NSDUHs was census tracts.6 This stage was included to contain sample segments within a single census tract to the extent possible.7

For each SS region, 48 census tracts were selected with probability proportional to size. Within sampled census tracts, adjacent census blocks were combined to form the second-stage sampling units or area segments. One area segment was selected within each sampled census tract with probability proportional to population size to support the 5-year sample and any supplemental studies that the Substance Abuse and Mental Health Services Administration (SAMHSA) may choose to field.8 Of these segments, 24 were designated for the coordinated 5-year sample and 24 were designated as "reserve" segments. Eight sample segments per SS region were fielded during the 2007 survey year.

These sampled segments were allocated equally into four separate samples, one for each 3-month period (calendar quarter) during the year. That is, a sample was selected from two segments in each calendar quarter so that the survey was essentially continuous in the field. In each of the area segments, a listing of all addresses was made from which a national sample of 192,092 addresses was selected. Of the selected addresses, 158,411 were determined to be eligible sample units. In these sample units (which can be either households or units within group quarters), sample persons were randomly selected using an automated screening procedure programmed in a handheld computer carried by the interviewers. The number of sample units completing the screening was 141,487. Youths aged 12 to 17 years and young adults aged 18 to 25 years were oversampled at this stage, with 12 to 17 year olds sampled at a rate of 85.2 percent and 18 to 25 year olds at a rate of 75.5 percent on average, when they were present in the sampled households or group quarters. Persons in age groups 26 or older were sampled at rates of 22.1 percent or less, with persons in the eldest age group (50 years or older) sampled at a rate of 8.2 percent on average. The overall population sampling rates were 0.09 percent for 12 to 17 year olds, 0.07 percent for 18 to 25 year olds, 0.02 percent for 26 to 34 year olds, 0.02 percent for 35 to 49 year olds, and 0.01 percent for those 50 or older. Because of the large sample size, there was no need to oversample racial/ethnic groups, as was done on surveys prior to 1999. Nationwide, 85,774 persons were selected. Consistent with previous surveys in this series, the final respondent sample of 67,870 persons was representative of the U.S. general population (since 1991, the civilian, noninstitutionalized population) aged 12 or older. In addition, State samples were representative of their respective State populations. More detailed information on the disposition of the national screening and interview sample can be found in Appendix B.

The survey covers residents of households (living in houses/townhouses, apartments, condominiums, etc.), persons in noninstitutional group quarters (e.g., shelters, rooming/boarding houses, college dormitories, migratory workers' camps, halfway houses), and civilians living on military bases. Although the survey covers residents of these types of units (they are given a nonzero probability of selection), the sample sizes of most specific groups are too small to provide separate estimates.

More information on the sample design can be found in the 2007 NSDUH sample design report by Morton et al. (2008) on the Office of Applied Studies (OAS) website (available as a PDF at http://oas.samhsa.gov/nsduh/methods.cfm).

A.2 Data Collection Methodology

The data collection method used in NSDUH involves in-person interviews with sample persons, incorporating procedures that would be likely to increase respondents' cooperation and willingness to report honestly about their illicit drug use behavior. Confidentiality is stressed in all written and oral communications with potential respondents. Respondents' names are not collected with the data, and computer-assisted interviewing (CAI) methods are used to provide a private and confidential setting to complete the interview.

Introductory letters are sent to sampled addresses, followed by an interviewer visit. A 5-minute screening procedure using a handheld computer involves listing all household members along with their basic demographic data. The computer uses the demographic data in a preprogrammed selection algorithm to select zero to two sample persons, depending on the composition of the household. This selection process is designed to provide the necessary sample sizes for the specified population age groupings. In areas where a third or more of the households contain Spanish-speaking residents, the initial introductory letters written in English are mailed with a Spanish version on the back. All interviewers carry copies of this letter in Spanish. If the interviewer is not certified bilingual, he or she will use preprinted Spanish cards to attempt to find someone in the household who speaks English and who can serve as the screening respondent or who can translate for the screening respondent. If no one is available, the interviewer will schedule a time when a Spanish-speaking interviewer can come to the address. In households where a language other than Spanish is encountered, another language card is used to attempt to find someone who speaks English to complete the screening.

The NSDUH interview is available in English and Spanish, and both versions have the same content. If the sample person prefers to complete the interview in Spanish, a certified bilingual interviewer is sent to the address to conduct the interview. Because the interview is not translated into any other language, if a sample person does not speak English or Spanish, the interview is not conducted.

Interviewers attempt to conduct the NSDUH interview immediately with each sample person in the household. The interviewer requests the selected respondent to identify a private area in the home to conduct the interview away from other household members. The interview averages about an hour and includes a combination of CAPI (computer-assisted personal interviewing, in which the interviewer reads the questions) and ACASI (audio computer-assisted self-interviewing).

The NSDUH interview consists of core and noncore (i.e., supplemental) sections. A core set of questions critical for basic trend measurement of prevalence estimates remains in the survey every year and comprises the first part of the interview. Noncore questions, or modules, that can be revised, dropped, or added from year to year make up the remainder of the interview. The core consists of initial demographic items (which are interviewer-administered) and self-administered questions pertaining to the use of tobacco, alcohol, marijuana, cocaine, crack cocaine, heroin, hallucinogens, inhalants, pain relievers, tranquilizers, stimulants, and sedatives. Topics in the remaining noncore self-administered sections include (but are not limited to) injection drug use, perceived risks of substance use, substance dependence or abuse, arrests, treatment for substance use problems, pregnancy and health care issues, and mental health issues. Noncore demographic questions (which are interviewer-administered and follow the ACASI questions) address such topics as immigration, current school enrollment, employment and workplace issues, health insurance coverage, and income. It should be noted that some of the noncore portions of the interview have remained in the survey, relatively unchanged, from year to year (e.g., current health insurance coverage, employment).

Thus, the interview begins in CAPI mode with the field interviewer (FI) reading the questions from the computer screen and entering the respondent's replies into the computer. The interview then transitions to the ACASI mode for the sensitive questions. In this mode, the respondent can read the questions silently on the computer screen and/or listen to the questions read through headphones and enter his or her responses directly into the computer. At the conclusion of the ACASI section, the interview returns to the CAPI mode with the interviewer completing the questionnaire. Each respondent who completes a full interview is given a $30.00 cash payment as a token of appreciation for his or her time.

No personal identifying information is captured in the CAI record for the respondent. Interviewers transmit the completed interview data to RTI in Research Triangle Park, North Carolina, via home telephone lines.

A.3 Data Processing

Computers at RTI direct the information to a raw data file (i.e., in which no logical editing of the data had been done) that consists of one record for each completed interview. Cases are retained only if respondents provided data on lifetime use of cigarettes and at least nine other substances in the core section of the questionnaire. Written responses to questions (e.g., names of other drugs that were used) are assigned numeric codes as part of the data processing procedures. Even though editing and consistency checks are done by the CAI program during the interview, additional, more complex edits and consistency checks are completed at RTI. Additionally, statistical imputation is used to replace missing or ambiguous values after editing for some key variables. Analysis weights are created so that estimates will be representative of the target population.

A.3.1 Data Coding and Logical Editing

With the exception of industry and occupation data, coding of written answers that respondents or interviewers typed was performed at RTI for the 2007 NSDUH. These written answers include mentions of drugs that respondents had used or other responses that did not fit a previous response option (subsequently referred to as "OTHER, Specify" data). Coding of the "OTHER, Specify" variables was accomplished through computer-assisted survey procedures and the use of a secure website that allowed for coding and review of the data. The computer-assisted procedures entailed a database check for a given "OTHER, Specify" variable that contained typed entries and the associated numeric codes. If an exact match was found between the typed response and an entry in the system, the computer-assisted procedures assigned the appropriate numeric code. Typed responses that did not match an existing entry were coded through the web-based coding system. Data on the industries in which respondents worked and respondents' occupations were assigned numeric industry and occupation codes by staff at the U.S. Census Bureau.

As noted above, the CAI program included checks that alerted respondents or interviewers when an entered answer was inconsistent with a previous answer in a given module. In this way, the inconsistency could be resolved while the interview was in progress. However, not every inconsistency was resolved during the interview, and the CAI program did not include checks for every possible inconsistency that might have occurred in the data.

Therefore, the first important step in processing the raw NSDUH data was logical editing of the data. Logical editing involved using data from within a respondent's record to (a) reduce the amount of item nonresponse (i.e., missing data) in interview records, including identification of items that were legitimately skipped; (b) make related data elements consistent with each other; and (c) identify ambiguities or inconsistencies to be resolved through statistical imputation procedures (see Section A.3.2).

For example, if respondents reported that they never used a given drug, the CAI logic skipped them out of all remaining questions about use of that drug. In the editing procedures, the skipped variables were assigned codes to indicate that the respondents were lifetime nonusers. Similarly, respondents were instructed in the prescription psychotherapeutics modules (i.e., pain relievers, tranquilizers, stimulants, and sedatives) not to report the use of over-the-counter (OTC) drugs. Therefore, if a respondent's only report of lifetime use of a particular type of "prescription" psychotherapeutic drug was for an OTC drug, the respondent was logically inferred never to have been a nonmedical user of the prescription drugs in that psychotherapeutic category.

In addition, respondents could report that they were lifetime users of a drug but not provide specific information on when they last used it. In this situation, a temporary "indefinite" value for the most recent period of use was assigned to the edited recency-of-use variable (e.g., Used at some point in the lifetime LOGICALLY ASSIGNED), and a final, specific value was statistically imputed. The editing procedures for key drug use variables also involved identifying inconsistencies between related variables so that these inconsistencies could be resolved through statistical imputation. For example, if a respondent reported last using a drug more than 12 months ago and also reported first using it at his or her current age, both of those responses could not be true. In this example, the inconsistent period of most recent use was replaced with an "indefinite" value, and the inconsistent age at first use was replaced with a missing data code. These indefinite or missing values were subsequently imputed through statistical procedures to yield consistent data for the related measures, as discussed in the next section.

A.3.2 Statistical Imputation

For some key variables that still had missing or ambiguous values after editing, statistical imputation was used to replace these values with appropriate response codes. For example, a response is ambiguous if the editing procedures assigned a respondent's most recent use of a drug to "use at some point in the lifetime," with no definite period within the lifetime. In this case, the imputation procedures assign a definite value for when the respondent last used the drug (e.g., in the past 30 days, more than 30 days ago but within the past 12 months, more than 12 months ago). Similarly, if a response is completely missing, the imputation procedures replace missing values with nonmissing ones.

In most cases, missing or ambiguous values are imputed in NSDUH using a methodology called predictive mean neighborhoods (PMN), which was developed specifically for the 1999 survey and used in all subsequent survey years. The PMN method offers a rigorous and flexible method that was implemented to improve the quality of estimates and allow more variables to be imputed. Some of the key reasons for implementing this method include the following: (1) the ability to use covariates to determine donors is far greater than that offered in the hot deck, (2) the relative importance of covariates can be determined by standard estimating equation techniques, (3) the correlations across response variables can be accounted for by making the imputation multivariate, and (4) sampling weights can be easily incorporated in the models. The PMN method has some similarity with the predictive mean matching method of Rubin (1986) except that, for the donor records, Rubin used the observed variable value (not the predictive mean) to compute the distance function. Also, the well-known method of nearest neighbor imputation is similar to PMN, except that the distance function is in terms of the original predictor variables and often requires somewhat arbitrary scaling of discrete variables. PMN is a combination of a model-assisted imputation methodology and a random nearest neighbor hot-deck procedure. The hot-deck procedure is set up in such a way that imputed values are made consistent with preexisting nonmissing values for other variables. Whenever feasible, the imputation of variables using PMN is multivariate, in which imputation is accomplished on several response variables at once. Variables requiring imputation using PMN are the core demographic variables, core drug use variables (recency of use, frequency of use, and age at first use), income, health insurance, and noncore demographic variables for work status, immigrant status, and the household roster. A weighted regression imputation is used to impute some of the missing values in the nicotine dependence variables.

In the modeling stage of PMN, the model chosen depends on the nature of the response variable Y. In the 2007 NSDUH, the models included binomial logistic regression, multinomial logistic regression, Poisson regression, and ordinary linear regression, where the models incorporated the sampling design weights.

In general, hot-deck imputation replaces an item nonresponse (missing or ambiguous value) with a recorded response that is donated from a "similar" respondent who has nonmissing data. For random nearest neighbor hot-deck imputation, the missing or ambiguous value is replaced by a responding value from a donor randomly selected from a set of potential donors. Potential donors are those defined to be "close" to the unit with the missing or ambiguous value according to a predefined function called a distance metric. In the hot-deck stage of PMN, the set of candidate donors (the "neighborhood") consists of respondents with complete data who have a predicted mean close to that of the item nonrespondent. The predicted means are computed both for respondents with and without missing data, which differs from Rubin's method where predicted means are not computed for the donor respondent (Rubin, 1986). In particular, the neighborhood consists of either the set of the closest 30 respondents or the set of respondents with a predicted mean (or means) within 5.0 percent of the predicted mean(s) of the item nonrespondent, whichever set is smaller. If no respondents are available who have a predicted mean (or means) within 5.0 percent of the item nonrespondent, the respondent with the predicted mean(s) closest to that of the item nonrespondent is selected as the donor.

In the univariate case (where only one variable is imputed using PMN), the neighborhood of potential donors is determined by calculating the relative distance between the predicted mean for an item nonrespondent and the predicted mean for each potential donor, then choosing those means defined by the distance metric. The pool of donors is restricted further to satisfy logical constraints whenever necessary (e.g., age at first crack use must not be less than age at first cocaine use).

Whenever possible, missing or ambiguous values for more than one response variable are considered at a time. In this (multivariate) case, the distance metric is a Mahalanobis distance (Manly, 1986) rather than a relative Euclidean distance. Whether the imputation is univariate or multivariate, only missing or ambiguous values are replaced, and donors are restricted to be logically consistent with the response variables that are not missing. Furthermore, donors are restricted to satisfy "likeness constraints" whenever possible. That is, donors are required to have the same values for variables highly correlated with the response. If no donors are available who meet these conditions, these likeness constraints can be loosened. For example, donors for the age at first use variable are required to be of the same age as recipients, if at all possible. Further details on the PMN methodology are provided in RTI International (2008) and by Singh, Grau, and Folsom (2001, 2002).

Although statistical imputation could not proceed separately within each State due to insufficient pools of donors, information about each respondent's State of residence was incorporated in the modeling and hot-deck steps. For most drugs, respondents were separated into three "State usage" categories as follows: respondents from States with high usage of a given drug were placed in one category, respondents from States with medium usage into another, and the remainder into a third category. This categorical "State rank" variable was used as one set of covariates in the imputation models. In addition, eligible donors for each item nonrespondent were restricted to be of the same State usage category (i.e., the same "State rank") as the nonrespondent.

A.3.3 Development of Analysis Weights

The general approach to developing and calibrating analysis weights involved developing design-based weights, dk, as the product of the inverse of the selection probabilities at each selection stage. Similar to the 2005 and 2006 NSDUHs, the 2007 NSDUH used a four-stage sample selection scheme in which an extra selection stage of census tracts was added before the selection of a segment. Thus, the design-based weights, dk, for the 2007 NSDUH incorporated an extra layer of sampling selection to reflect the sample design change. Adjustment factors, ak(λ), then were applied to the design-based weights to adjust for nonresponse, to poststratify to known population control totals, and to control for extreme weights when necessary. In view of the importance of State-level estimates with the 50-State design, it was necessary to control for a much larger number of known population totals. Several other modifications to the general weight adjustment strategy that had been used in past surveys also were implemented for the first time beginning with the 1999 CAI sample.

Weight adjustments were based on a generalization of Deville and Särndal's (1992) logit model. This generalized exponential model (GEM) (Folsom & Singh, 2000b) incorporates unit-specific bounds (image representing lower case script lk, uk), ks, for the adjustment factor ak(λ) as follows:

Appendix A Equation,     D



where ck are prespecified centering constants, such that image representing lower case script lk < ck < uk and Ak = (uk - image representing lower case script lk) / (uk - ck)(ck - image representing lower case script lk). The variables image representing lower case script lk, ck, and uk are user-specified bounds, and λ is the column vector of p model parameters corresponding to the p covariates x. The λ-parameters are estimated by solving

Appendix A Equation,     D

where image representing uppercase T topped by a tildex denotes control totals that could be either nonrandom, as is generally the case with poststratification, or random, as is generally the case for nonresponse adjustment.

The final weights wk = dkak(λ) minimize the distance function Δ(w,d) defined as

Appendix A Equation,     D



This general approach was used at several stages of the weight adjustment process, including (1) adjustment of household weights for nonresponse at the screener level, (2) poststratification of household weights to meet population controls for various demographic groups by State, (3) adjustment of household weights for extremes, (4) poststratification of selected person weights, (5) adjustment of responding person weights for nonresponse at the questionnaire level, (6) poststratification of responding person weights, and (7) adjustment of responding person weights for extremes.

Every effort was made to include as many relevant State-specific covariates (typically defined by demographic domains within States) as possible in the multivariate models used to calibrate the weights (nonresponse adjustment and poststratification steps). Because further subdivision of State samples by demographic covariates often produced small cell sample sizes, it was not possible to retain all State-specific covariates (even after meaningful collapsing of covariate categories) and still estimate the necessary model parameters with reasonable precision. Therefore, a hierarchical structure was used in grouping States with covariates defined at the national level, at the census division level within the Nation, at the State group within the census division, and, whenever possible, at the State level. In every case, the controls for the total population within a State and the five age groups (12 to 17, 18 to 25, 26 to 34, 35 to 49, 50 or older) within a State were maintained except that, in the last step of poststratification of person weights, six age groups (12 to 17, 18 to 25, 26 to 34, 35 to 49, 50 to 64, 65 or older) were used. Census control totals by age, race, gender, and Hispanicity were required for the civilian, noninstitutionalized population of each State. Beginning with the 2002 NSDUH, the Population Estimates Branch of the U.S. Census Bureau has produced the necessary population estimates in response to a special request based on the 2000 census.

Consistent with the surveys from 1999 onward, control of extreme weights through separate bounds for adjustment factors was incorporated into the GEM calibration processes for both nonresponse and poststratification. This is unlike the traditional method of winsorization in which extreme weights are truncated at prespecified levels and the trimmed portions of weights are distributed to the nontruncated cases. In GEM, it is possible to set bounds around the prespecified levels for extreme weights, and then the calibration process provides an objective way of deciding the extent of adjustment (or truncation) within the specified bounds. A step was added to poststratify the household-level weights to obtain census-consistent estimates based on the household rosters from all screened households; these household roster-based estimates then provided the control totals needed to calibrate the respondent pair weights for subsequent planned analyses. An additional step poststratified the selected person sample to conform to the adjusted roster estimates. This additional step takes advantage of the inherent two-phase nature of the NSDUH design. The final step poststratified the respondent person sample to external census data (defined within the State whenever possible, as discussed above). For more detailed information, see the 2006 NSDUH Methodological Resource Book (RTI International, 2008).

For certain populations of interest, 2 years of NSDUH data were combined to obtain annual averages. The person-level weights for estimates based on the annual averages were obtained by dividing the analysis weights for the 2 specific years by a factor of 2.


End Notes

3 Prior to 2002, the survey was known as the National Household Survey on Drug Abuse (NHSDA).

4 SAE is a hierarchical Bayes modeling technique used to make State-level estimates for approximately 20 measures related to substance use. For more details, see the State Estimates of Substance Use from the 2005-2006 National Surveys on Drug Use and Health (Hughes, Sathe, & Spagnola, 2008).

5 Areas were defined using 2000 census geography. Dwelling units (DUs) and population counts were obtained from the 2000 census data supplemented with revised population counts from Claritas (http://www.claritas.com/Default.jsp).

6 Census tracts are relatively permanent statistical subdivisions of counties and provide a stable set of geographic units across decennial census periods.

7 Some census tracts had to be aggregated in order to meet the minimum DU requirement of 150 DUs in urban areas and 100 DUs in rural areas.

8 For more details on the 5-year sample, see the 2007 sample design report in the 2007 NSDUH Methodological Resource Book (Morton, Martin, Hirsch, & Chromy, 2008).

Go to Top of PageGo to the Table of Contents

This is the page footer.

This page was last updated on December 30, 2008.

SAMHSA, an agency in the Department of Health and Human Services, is the Federal Government's lead agency for improving the quality and availability of substance abuse prevention, addiction treatment, and mental health services in the United States.

This is a line.

   Site Map | Contact Us | AccessibilityPrivacy PolicyFreedom of Information Act
 Disclaimer | Department of Health and Human ServicesSAMHSAWhite HouseUSA.gov

* PDF formatted files require that Adobe Acrobat Reader® program is installed on your computer. Click here to download this FREE software now from Adobe.