Causal analysis is the field of experimental design and statistics pertaining to establishing cause and effect.[1] Typically it involves establishing four elements: correlation, sequence in time (that is, causes must occur before their proposed effect), a plausible physical or information-theoretical mechanism for an observed effect to follow from a possible cause, and eliminating the possibility of common and alternative ("special") causes. Such analysis usually involves one or more artificial or natural experiments.[2]

Motivation

Data analysis is primarily concerned with causal questions.[3][4][5][6][7] For example, did the fertilizer cause the crops to grow?[8] Or, can a given sickness be prevented?[9] Or, why is my friend depressed?[10] The potential outcomes and regression analysis techniques handle such queries when data is collected using designed experiments. Data collected in observational studies require different techniques for causal inference (because, for example, of issues such as confounding).[11] Causal inference techniques used with experimental data require additional assumptions to produce reasonable inferences with observation data.[12] The difficulty of causal inference under such circumstances is often summed up as "correlation does not imply causation".

In philosophy and physics

The nature of causality is systematically investigated in several academic disciplines, including philosophy and physics.

In academia, there are a significant number of theories on causality; The Oxford Handbook of Causation (Beebee, Hitchcock & Menzies 2009) encompasses 770 pages. Among the more influential theories within philosophy are Aristotle's Four causes and Al-Ghazali's occasionalism.[13] David Hume argued that beliefs about causality are based on experience, and experience similarly based on the assumption that the future models the past, which in turn can only be based on experience – leading to circular logic. In conclusion, he asserted that causality is not based on actual reasoning: only correlation can actually be perceived.[14] Immanuel Kant, according to Beebee, Hitchcock & Menzies (2009), held that "a causal principle according to which every event has a cause, or follows according to a causal law, cannot be established through induction as a purely empirical claim, since it would then lack strict universality, or necessity".

Outside the field of philosophy, theories of causation can be identified in classical mechanics, statistical mechanics, quantum mechanics, spacetime theories, biology, social sciences, and law.[13] To establish a correlation as causal within physics, it is normally understood that the cause and the effect must connect through a local mechanism (cf. for instance the concept of impact) or a nonlocal mechanism (cf. the concept of field), in accordance with known laws of nature.

From the point of view of thermodynamics, universal properties of causes as compared to effects have been identified through the Second law of thermodynamics, confirming the ancient, medieval and Cartesian[15] view that "the cause is greater than the effect" for the particular case of thermodynamic free energy. This, in turn, is challenged by popular interpretations of the concepts of nonlinear systems and the butterfly effect, in which small events cause large effects due to, respectively, unpredictability and an unlikely triggering of large amounts of potential energy.

Causality construed from counterfactual states

Intuitively, causation seems to require not just a correlation, but a counterfactual dependence. Suppose that a student performed poorly on a test and guesses that the cause was his not studying. To prove this, one thinks of the counterfactual – the same student writing the same test under the same circumstances but having studied the night before. If one could rewind history, and change only one small thing (making the student study for the exam), then causation could be observed (by comparing version 1 to version 2). Because one cannot rewind history and replay events after making small controlled changes, causation can only be inferred, never exactly known. This is referred to as the Fundamental Problem of Causal Inference – it is impossible to directly observe causal effects.[16]

A major goal of scientific experiments and statistical methods is to approximate as best possible the counterfactual state of the world.[17] For example, one could run an experiment on identical twins who were known to consistently get the same grades on their tests. One twin is sent to study for six hours while the other is sent to the amusement park. If their test scores suddenly diverged by a large degree, this would be strong evidence that studying (or going to the amusement park) had a causal effect on test scores. In this case, correlation between studying and test scores would almost certainly imply causation.

Well-designed experimental studies replace equality of individuals as in the previous example by equality of groups. The objective is to construct two groups that are similar except for the treatment that the groups receive. This is achieved by selecting subjects from a single population and randomly assigning them to two or more groups. The likelihood of the groups behaving similarly to one another (on average) rises with the number of subjects in each group. If the groups are essentially equivalent except for the treatment they receive, and a difference in the outcome for the groups is observed, then this constitutes evidence that the treatment is responsible for the outcome, or in other words the treatment causes the observed effect. However, an observed effect could also be caused "by chance", for example as a result of random perturbations in the population. Statistical tests exist to quantify the likelihood of erroneously concluding that an observed difference exists when in fact it does not (for example see P-value).

Operational definitions of causality

Clive Granger created the first operational definition of causality in 1969.[18] Granger made the definition of probabilistic causality proposed by Norbert Wiener operational as a comparison of variances.[19]

Verification by "truth"

Peter Spirtes, Clark Glymour, and Richard Scheines introduced the idea of explicitly not providing a definition of causality .[3] Spirtes and Glymour introduced the PC algorithm for causal discovery in 1990.[20] Many recent causal discovery algorithms follow the Spirtes-Glymour approach to verification.[21]

Exploratory

Exploratory causal analysis, also known as "data causality" or "causal discovery"[3] is the use of statistical algorithms to infer associations in observed data sets that are potentially causal under strict assumptions. ECA is a type of causal inference distinct from causal modeling and treatment effects in randomized controlled trials.[4] It is exploratory research usually preceding more formal causal research in the same way exploratory data analysis often precedes statistical hypothesis testing in data analysis.[22][23]

See also

References

  1. Rohlfing, Ingo; Schneider, Carsten Q. (2018). "A Unifying Framework for Causal Analysis in Set-Theoretic Multimethod Research" (PDF). Sociological Methods & Research. 47 (1): 37–63. doi:10.1177/0049124115626170. S2CID 124804330. Retrieved 29 February 2020.
  2. Brady, Henry E. (7 July 2011). "Causation and Explanation in Social Science". The Oxford Handbook of Political Science. doi:10.1093/oxfordhb/9780199604456.013.0049. Retrieved 29 February 2020.
  3. 1 2 3 Spirtes, P.; Glymour, C.; Scheines, R. (2012). Causation, Prediction, and Search. Springer Science & Business Media. ISBN 978-1461227489.
  4. 1 2 Rosenbaum, Paul (2017). Observation and Experiment: An Introduction to Causal Inference. Harvard University Press. ISBN 9780674975576.
  5. Pearl, Judea; Mackenzie, Dana (2018). The Book of Why: The New Science of Cause and Effect. Basic Books. ISBN 978-0465097616.
  6. Kleinberg, Samantha (2015). Why: A Guide to Finding and Using Causes. O'Reilly Media, Inc. ISBN 978-1491952191.
  7. Illari, P.; Russo, F. (2014). Causality: Philosophical Theory meets Scientific Practice. OUP Oxford. ISBN 978-0191639685.
  8. Fisher, R. (1937). The design of experiments. Oliver And Boyd.
  9. Hill, B. (1955). Principles of Medical Statistics. Lancet Limited.
  10. Halpern, J. (2016). Actual Causality. MIT Press. ISBN 978-0262035026.
  11. Pearl, J.; Glymour, M.; Jewell, N. P. (2016). Causal inference in statistics: a primer. John Wiley & Sons. ISBN 978-1119186847.
  12. Stone, R. (1993). "The Assumptions on Which Causal Inferences Rest". Journal of the Royal Statistical Society. Series B (Methodological). 55 (2): 455–466. doi:10.1111/j.2517-6161.1993.tb01915.x.
  13. 1 2 Beebee, Hitchcock & Menzies 2009
  14. Morris, William Edward (2001). "David Hume". The Stanford Encyclopedia of Philosophy.
  15. Lloyd, A.C. (1976). "The principle that the cause is greater than its effect". Phronesis. 21 (2): 146–156. doi:10.1163/156852876x00101. JSTOR 4181986.
  16. Holland, Paul W. (1986). "Statistics and Causal Inference". Journal of the American Statistical Association. 81 (396): 945–960. doi:10.1080/01621459.1986.10478354. S2CID 14377504.
  17. Pearl, Judea (2000). Causality: Models, Reasoning, and Inference. Cambridge University Press. ISBN 9780521773621.
  18. Granger, C. W. J. (1969). "Investigating Causal Relations by Econometric Models and Cross-spectral Methods". Econometrica. 37 (3): 424–438. doi:10.2307/1912791. JSTOR 1912791.
  19. Granger, Clive. "Prize Lecture. NobelPrize.org. Nobel Media AB 2018".
  20. Spirtes, P.; Glymour, C. (1991). "An algorithm for fast recovery of sparse causal graphs". Social Science Computer Review. 9 (1): 62–72. doi:10.1177/089443939100900106. S2CID 38398322.
  21. Guo, Ruocheng; Cheng, Lu; Li, Jundong; Hahn, P. Richard; Liu, Huan (2020). "A Survey of Learning Causality with Data". ACM Computing Surveys. 53 (4): 1–37. arXiv:1809.09337. doi:10.1145/3397269. S2CID 52822969.
  22. McCracken, James (2016). Exploratory Causal Analysis with Time Series Data (Synthesis Lectures on Data Mining and Knowledge Discovery). Morgan & Claypool Publishers. ISBN 978-1627059343.
  23. Tukey, John W. (1977). Exploratory Data Analysis. Pearson. ISBN 978-0201076165.

Bibliography

This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.