www.fgks.org   »   [go: up one dir, main page]

Confounding: Difference between revisions

Content deleted Content added
m Added to the Simple Example
 
(19 intermediate revisions by 16 users not shown)
Line 1:
{{short description|Variable or factor in statisticscausal inference}}
{{Redirect|Confounding factor|the company|Confounding Factor|the psychological state|Confusion}}
 
{{technical|date=September 2019}}
 
[[File:comparison_confounder_mediatorcomparison confounder mediator.svg|thumb|upright=1.3|Whereas a mediator is a factor in the causal chain (above), a confounder is a spurious factor incorrectly implying causation (bottom)]]
 
In statistics[[causal inference]], a '''confounder''' (also '''confounding variable''', '''confounding factor''', '''extraneous determinant''' or '''lurking variable''') is a variable that influences both the [[dependent and independent variables|dependent variable and independent variable]], causing a [[Spurious relationship|spurious association]]. Confounding is a [[Causality|causal]] concept, and as such, cannot be described in terms of correlations or associations.<ref name="Pearl 2009">Pearl, J., (2009). [[Simpson's Paradox]], Confounding, and Collapsibility In ''Causality: Models, Reasoning and Inference'' (2nd ed.). New York : Cambridge University Press.</ref><ref>{{cite journal | last1 = VanderWeele | first1 = T.J. | last2 = Shpitser | first2 = I. | year = 2013 | title = On the definition of a confounder | journal = Annals of Statistics | volume = 41 | issue = 1| pages = 196–220 | doi=10.1214/12-aos1058| pmid = 25544784 | pmc = 4276366 | arxiv = 1304.0564 }}</ref><ref name="Greenland Pearl Robbins 1999">{{cite journal | last1 = Greenland | first1 = S. | last2 = Robins | first2 = J. M. | last3 = Pearl | first3 = J. | year = 1999 | title = Confounding and Collapsibility in Causal Inference | doi = 10.1214/ss/1009211805 | journal = Statistical Science | volume = 14 | issue = 1| pages = 29–46 | doi-access = free }}</ref> The existence of confounders is an important quantitative explanation why [[correlation does not imply causation]]. Some [[Causal notation|notations]] are explicitly designed to identify the existence, possible existence, or non-existence of confounders in causal relationships between elements of a system.
 
Confounds are threats to [[internal validity]].<ref name=Shadish2002>{{cite book|last1=Shadish|first1=W. R.|last2=Cook|first2=T. D.|last3=Campbell|first3=D. T.|year=2002|title=Experimental and quasi-experimental designs for generalized causal inference|location=Boston, MA|publisher=[[Houghton Mifflin|Houghton-Mifflin]]}}</ref>
 
== Simple Example ==
Let's assume that a trucking company owns a fleet of trucks made by two different manufacturers. Trucks made by one manufacturer are called "A Trucks" and trucks made by the other manufacturer are called "B Trucks." We want to find out whether A Trucks or B Trucks get better fuel economy. We measure fuel and miles driven for a month and calculate the MPG for each truck. We then run the appropriate analysis, which determines that there is a statistically significant trend that A Trucks are more fuel efficient than B Trucks. Upon further reflection, however, we also notice that A Trucks are more likely to be assigned highway routes, and B Trucks are more likely to be assigned city routes. This is a confounding variable. The confounding variable makes the results of the analysis unreliable. It is quite likely that we are just measuring the fact that highway driving results in better fuel economy than city driving.
 
In statistics terms, the make of the truck is the independent variable, the fuel economy (MPG) is the dependent variable and the amount of city driving is the confounding variable. To fix this study, we have several choices. One is to randomize the truck assignments so that A trucks and B Trucks end up with equal amounts of city and highway driving. That eliminates the confounding variable. Another choice is to quantify the amount of city driving and use that as a second independent variable. A third choice is to segment the study, first comparing MPG during city driving for all trucks, and then run a separate study comparing MPG during highway driving.
 
== Definition ==
Line 31 ⟶ 36:
{{NumBlk|:|<math>P(y \mid \text{do}(x)) \ne P(y \mid x)</math>|{{EquationRef|2}}}}
 
because the observational quantity contains information about the correlation between ''X'' and ''Z'', and the interventional quantity does not (since ''X'' is not correlated with ''Z'' in a randomized experiment). It can be shown <ref name="Pearl 1993">Pearl, J., (1993). "Aspects of Graphical Models Connected With Causality,", ''In Proceedings of the 49th Session of the International Statistical Science Institute,'' pp. 391–401.</ref> that, in cases where only observational data areis available, an unbiased estimate of the desired quantity <math>P(y \mid \text{do}(x))</math>, can
be obtained by "adjusting" for all confounding factors, namely, conditioning on their various values and averaging the result. In the case of a single confounder ''Z'', this leads to the "adjustment formula":
 
{{NumBlk|:|<math>P(y \mid \text{do}(x)) = \sum_{z} P(y \mid x, z) P(z)</math>|{{EquationRef|3}}}}
 
which gives an unbiased estimate for the causal effect of ''X'' on ''Y''. The same adjustment formula works when there are multiple confounders except, in this case, the choice of a set ''Z'' of variables that would guarantee unbiased estimates must be done with caution. The criterion for a proper choice of variables is called the Back-Door <ref name="Pearl 1993"></ref><ref name="pearl09-causal-diagrams">Pearl, J. (2009). Causal Diagrams and the Identification of Causal Effects In ''Causality: Models, Reasoning and Inference'' (2nd ed.). New York, NY, USAUS: Cambridge University Press.</ref> and requires that the chosen set ''Z'' "blocks" (or intercepts) every path between ''X'' and ''Y'' that contains an arrow into X. Such sets are called "Back-Door admissible" and may include variables which are not common causes of ''X'' and ''Y'', but merely proxies thereof.
 
Returning to the drug use example, since ''Z'' complies with the Back-Door requirement (i.e., it intercepts the one Back-Door path <math>X \leftarrow Z \rightarrow Y</math>), the Back-Door adjustment formula is valid:
Line 44 ⟶ 49:
In this way the physician can predict the likely effect of administering the drug from observational studies in which the conditional probabilities appearing on the right-hand side of the equation can be estimated by regression.
 
Contrary to common beliefs, adding covariates to the adjustment set ''Z'' can introduce bias.<ref>{{cite web |last1=Cinelli |first1=C.|last2=Forney |first2=A.|last3=Pearl |first3=J.|title=A Crash Course in Good and Bad Controls |url=http://ftp.cs.ucla.edu/pub/stat_ser/r493.pdf |website=UCLA Cognitive Systems Laboratory, Technical Report (R-493) |date=March 2022}}</ref> A typical counterexample occurs when ''Z'' is a common effect of ''X'' and ''Y'',<ref>{{cite journal |last=Lee |first=P. H. |title=Should We Adjust for a Confounder if Empirical and Theoretical Criteria Yield Contradictory Results? A Simulation Study |journal=[[Scientific Reports|Sci Rep]] |year=2014 |volume=4 |pages=6085 |doi=10.1038/srep06085 |pmid=25124526 |bibcode=2014NatSR...4E6085L |pmc=5381407 }}</ref> a case in which ''Z'' is not a confounder (i.e., the null set is Back-door admissible) and adjusting for ''Z'' would create bias known as "[[Collider (epidemiology)|collider]] bias" or "[[Berkson's paradox]]." Controls that are not good confounders are sometimes called [[bad control]]s.
 
In general, confounding can be controlled by adjustment if and only if there is a set of observed covariates that satisfies the Back-Door condition. Moreover, if ''Z'' is such a set, then the adjustment formula of Eq. (3) is valid.<ref name="Pearl 1993"></ref><ref name="pearl09-causal-diagrams">Pearl, J. (2009). Causal Diagrams and the Identification of Causal Effects In ''Causality: Models, Reasoning and Inference'' (2nd ed.). New York, NY, USAUS: Cambridge University Press.</ref> Pearl's do-calculus provides all possible conditions under which <math>P(y \mid \text{do}(x))</math> can be estimated, not necessarily by adjustment.<ref>{{cite journal | last1 = Shpitser | first1 = I. | last2 = Pearl | first2 = J. | year = 2008 | title = Complete identification methods for the causal hierarchy | journal = The Journal of Machine Learning Research | volume = 9 | pages = 1941–1979 }}</ref>
 
==History==
According to Morabia (2011),<ref>{{cite journal | last1 = Morabia | first1 = A | year = 2011 | title = History of the modern epidemiological concept of confounding | url =http://jech.bmj.com/content/jech/65/4/297.full.pdf | journal = Journal of Epidemiology and Community Health | volume = 65 | issue = 4| pages = 297–300 | doi=10.1136/jech.2010.112565| pmid = 20696848 | s2cid = 9068532 | doi-access = free }}</ref> the word ''confounding'' derives from the [[Medieval Latin]] verb "confudereconfundere", which meant "mixing", and was probably chosen to represent the confusion (from Latin: con=with + fusus=mix or fuse together) between the cause one wishes to assess and other causes that may affect the outcome and thus confuse, or stand in the way of the desired assessment. [[Ronald Fisher|Fisher]] used the word "confounding" in his 1935 book "The Design of Experiments"<ref>FisherGreenland, R.Robins A.and (1935). The design of experiments (pp. 114–145).Pearl</ref> toname="Greenland denoteRobins anyPearl source of error in his ideal of randomized experiment. According to Vandenbroucke (2004)<ref1999">{{cite journal | last1 = VandenbrouckeGreenland | first1 = J. PS. | yearlast2 = 2004Robins | titlefirst2 = TheJ. history of confoundingM. | journallast3 = Soz PraventivmedPearl | volumefirst3 = 47 | issue = 4| pages = 216–224 | doi = 10J.1007/BF01326402 | pmidyear = 124159251999 | s2cidtitle = 198174446Confounding }}</ref>and itCollapsibility wasin [[LeslieCausal Kish|Kish]]<ref>{{cite journalInference | last1doi = Kish | first1 = L | year = 1959 | title = Some statistical problems in research design10.1214/ss/1009211805 | journal = AmStatistical SociolScience | volume = 2614 | issue = 31| pagespage = 328–33831 | doi-access = 10.2307/2089381 | jstor = 2089381free }}</ref> whonote usedan theearly word "confounding" in the modern senseuse of the word, to meanterm "incomparabilityconfounding" ofin twocausal orinference moreby groupsJohn (e.g.,Stuart exposed and unexposed)Mill in an observational study1843.
 
[[Ronald Fisher|Fisher]] introduced the word "confounding" in his 1935 book "The Design of Experiments"<ref>Fisher, R. A. (1935). The design of experiments (pp. 114–145).</ref> to refer specifically to a consequence of [[Blocking (statistics)|blocking]] (i.e., [[Partition of a set|partitioning]]) the set of treatment combinations in a [[factorial experiment]], whereby certain interactions may be "confounded with blocks". This popularized the notion of confounding in statistics, although Fisher was concerned with the control of heterogeneity in experimental units, not with causal inference.
Formal conditions defining what makes certain groups "comparable" and others "incomparable" were later developed in [[epidemiology]] by Greenland and Robins (1986)<ref>{{cite journal | last1 = Greenland | first1 = S. | last2 = Robins | first2 = J. M. | year = 1986 | title = Identifiability, exchangeability, and epidemiological confounding | journal = International Journal of Epidemiology | volume = 15 | issue = 3| pages = 413–419 | doi=10.1093/ije/15.3.413| pmid = 3771081 | citeseerx = 10.1.1.157.6445 }}</ref> using the counterfactual language of [[Jerzy Neyman|Neyman]] (1935)<ref>Neyman, J., with cooperation of K. Iwaskiewics and St. Kolodziejczyk (1935). Statistical problems in agricultural experimentation (with discussion). ''Suppl J Roy Statist Soc Ser'' B 2 107-180.</ref> and [[Donald Rubin|Rubin]] (1974).<ref>{{cite journal | last1 = Rubin | first1 = D. B. | s2cid = 52832751 | year = 1974 | title = Estimating causal effects of treatments in randomized and nonrandomized studies | journal = Journal of Educational Psychology | volume = 66 | issue = 5| pages = 688–701 | doi=10.1037/h0037350}}</ref> These were later supplemented by graphical criteria such as the Back-Door condition ([[Judea Pearl|Pearl]] 1993; Greenland, Pearl and Robins, 1999).<ref name="Greenland Pearl Robbins 1999" /><ref name="Pearl 1993" />
 
According to Vandenbroucke (2004)<ref>{{cite journal | last1 = Vandenbroucke | first1 = J. P. | year = 2004 | title = The history of confounding | journal = Soz Praventivmed | volume = 47 | issue = 4| pages = 216–224 | doi = 10.1007/BF01326402 | pmid = 12415925 | s2cid = 198174446 }}</ref> it was [[Leslie Kish|Kish]]<ref>{{cite journal | last1 = Kish | first1 = L | year = 1959 | title = Some statistical problems in research design | journal = Am Sociol | volume = 26 | issue = 3| pages = 328–338 | doi = 10.2307/2089381 | jstor = 2089381 }}</ref> who used the word "confounding" in the sense of "incomparability" of two or more groups (e.g., exposed and unexposed) in an observational study. Formal conditions defining what makes certain groups "comparable" and others "incomparable" were later developed in [[epidemiology]] by Greenland and Robins (1986)<ref>{{cite journal | last1 = Greenland | first1 = S. | last2 = Robins | first2 = J. M. | year = 1986 | title = Identifiability, exchangeability, and epidemiological confounding | journal = International Journal of Epidemiology | volume = 15 | issue = 3| pages = 413–419 | doi=10.1093/ije/15.3.413| pmid = 3771081 | citeseerx = 10.1.1.157.6445 }}</ref> using the counterfactual language of [[Jerzy Neyman|Neyman]] (1935)<ref>Neyman, J., with cooperation of K. Iwaskiewics and St. Kolodziejczyk (1935). Statistical problems in agricultural experimentation (with discussion). ''Suppl J Roy Statist Soc Ser'' B 2 107-180.</ref> and [[Donald Rubin|Rubin]] (1974).<ref>{{cite journal | last1 = Rubin | first1 = D. B. | s2cid = 52832751 | year = 1974 | title = Estimating causal effects of treatments in randomized and nonrandomized studies | journal = Journal of Educational Psychology | volume = 66 | issue = 5| pages = 688–701 | doi=10.1037/h0037350}}</ref> These were later supplemented by graphical criteria such as the Back-Door condition ([[Judea Pearl|Pearl]] 1993; Greenland, PearlRobins and Robins,Pearl 1999).<ref name="Greenland PearlRobins RobbinsPearl 1999" /><ref name="Pearl 1993" />
Graphical criteria were shown to be formally equivalent to the counterfactual definition<ref>Pearl, J., (2009). ''Causality: Models, Reasoning and Inference'' (2nd ed.). New York, NY, USA: Cambridge University Press.</ref> but more transparent to researchers relying on process models.
 
Graphical criteria were shown to be formally equivalent to the counterfactual definition<ref>Pearl, J., (2009). ''Causality: Models, Reasoning and Inference'' (2nd ed.). New York, NY, USAUS: Cambridge University Press.</ref> but more transparent to researchers relying on process models.
 
==Types==
Line 61 ⟶ 68:
 
Confounding variables may also be categorised according to their source. The choice of measurement instrument (operational confound), situational characteristics (procedural confound), or inter-individual differences (person confound).
* An '''operational confounding''' can occur in both [[experiment]]al and non-experimental research designs. This type of confounding occurs when a measure designed to assess a particular construct inadvertently measures something else as well.<ref name=Pelham>{{cite book |last=Pelham |first=Brett |year=2006 |title=Conducting Research in Psychology |location=Belmont |publisher=Wadsworth |isbn=978-0-534-53294-9 }}</ref>
* A '''procedural confounding''' can occur in a laboratory experiment or a [[quasi-experiment]]. This type of confound occurs when the researcher mistakenly allows another variable to change along with the manipulated independent variable.<ref name=Pelham/>
* A '''person confounding''' occurs when two or more groups of units are analyzed together (e.g., workers from different occupations), despite varying according to one or more other (observed or unobserved) characteristics (e.g., gender).<ref>{{cite book |last1=Steg |first1=L. |last2=Buunk |first2=A. P. |last3=Rothengatter |first3=T. |year=2008 |title=Applied Social Psychology: Understanding and managing social problems |location=Cambridge, UK |publisher=Cambridge University Press |chapter=Chapter 4 }}</ref>
 
== Examples ==
Say one is studying the relation between birth order (1st child, 2nd child, etc.) and the presence of [[Down Syndrome]] in the child. In this scenario, maternal age would be a confounding variable:{{cn|date=April 2024}}
# Higher maternal age is directly associated with Down Syndrome in the child
# Higher maternal age is directly associated with Down Syndrome, regardless of birth order (a mother having her 1st vs 3rd child at age 50 confers the same risk)
Line 94 ⟶ 101:
| url = https://archive.org/details/epidemiologyinme00henn
}}</ref>
 
* [[Case-control study|Case-control studies]] assign confounders to both groups, cases and controls, equally. For example, if somebody wanted to study the cause of myocardial infarct and thinks that the age is a probable confounding variable, each 67-year-old infarct patient will be matched with a healthy 67-year-old "control" person. In case-control studies, matched variables most often are the age and sex. Drawback: Case-control studies are feasible only when it is easy to find controls, ''i.e.'' persons whose status vis-à-vis all known potential confounding factors is the same as that of the case's patient: Suppose a case-control study attempts to find the cause of a given disease in a person who is 1) 45 years old, 2) African-American, 3) from [[Alaska#Demographics|Alaska]], 4) an avid football player, 5) vegetarian, and 6) working in education. A theoretically perfect control would be a person who, in addition to not having the disease being investigated, matches all these characteristics and has no diseases that the patient does not also have—but finding such a control would be an enormous task.
* [[Cohort study|Cohort studies]]: A degree of matching is also possible and it is often done by only admitting certain age groups or a certain sex into the study population, creating a cohort of people who share similar characteristics and thus all cohorts are comparable in regard to the possible confounding variable. For example, if age and sex are thought to be confounders, only 40 to 50 years old males would be involved in a cohort study that would assess the myocardial infarct risk in cohorts that either are physically active or inactive. Drawback: In cohort studies, the overexclusion of input data may lead researchers to define too narrowly the set of similarly situated persons for whom they claim the study to be useful, such that other persons to whom the causal relationship does in fact apply may lose the opportunity to benefit from the study's recommendations. Similarly, "over-stratification" of input data within a study may reduce the sample size in a given stratum to the point where generalizations drawn by observing the members of that stratum alone are not [[Statistical significance|statistically significant]].
Line 107 ⟶ 113:
 
==Artifacts==
Artifacts are variables that should have been systematically varied, either within or across studies, but that waswere accidentally held constant. Artifacts are thus threats to [[external validity]]. Artifacts are factors that covary with the treatment and the outcome. Campbell and Stanley<ref name=C&S2006>{{cite book|last1=Campbell|first1=D. T.|last2=Stanley|first2=J. C.|year=1966|title=Experimental and quasi-experimental designs for research|location=Chicago|publisher=Rand McNally}}</ref> identify several artifacts. The major threats to internal validity are history, maturation, testing, instrumentation, [[Regression analysis|statistical regression]], selection, experimental mortality, and selection-history interactions.
 
One way to minimize the influence of artifacts is to use a pretest-posttest [[Experimental control|control group]] design. Within this design, "groups of people who are initially equivalent (at the pretest phase) are randomly assigned to receive the experimental treatment or a control condition and then assessed again after this differential experience (posttest phase)".<ref name=C&B2002>{{cite book|last1=Crano|first1=W. D.|last2=Brewer|first2=M. B.|year=2002|title=Principles and methods of [[social research]]|edition=2nd|location=Mahwah, NJ|publisher=[[Lawrence Erlbaum Associates]]|page=28}}</ref> Thus, any effects of artifacts are (ideally) equally distributed in participants in both the treatment and control conditions.
Line 116 ⟶ 122:
* {{annotated link|Epidemiological method}}
* {{annotated link|Simpson's paradox}}
* [[Omitted-variable bias]]
 
== References ==
Line 132 ⟶ 139:
|pages=[https://archive.org/details/handbookofresear00reis/page/3 3–16]
|location=New York
|publisher=[[Cambridge University Press]]}}|isbn=9780521551281
}}
* {{cite book
|last1=Smith|first1=E. R.
Line 146 ⟶ 154:
 
==External links==
{{Selfref|These sites contain descriptions or examples of confounding variables:.}}
* [http://sphweb.bumc.bu.edu/otlt/MPH-Modules/BS/BS704-EP713_Confounding-EM/BS704-EP713_Confounding-EM_print.html Tutorial: Confounding and Effect Measure Modification (Boston University School of Public Health)]
* [http://www.stat.yale.edu/Courses/1997-98/101/linreg.htm Linear Regression (Yale University)]
Line 152 ⟶ 160:
 
{{statistics}}
{{Experimental design |state = collapsed }}
 
[[Category:Analysis of variance]]