Schweitzer Fachinformationen
Wenn es um professionelles Wissen geht, ist Schweitzer Fachinformationen wegweisend. Kunden aus Recht und Beratung sowie Unternehmen, öffentliche Verwaltungen und Bibliotheken erhalten komplette Lösungen zum Beschaffen, Verwalten und Nutzen von digitalen und gedruckten Medien.
The False Discovery Rate
An essential tool for statisticians and data scientists seeking to interpret the vast troves of data that increasingly power our world
First developed in the 1990s, the False Discovery Rate (FDR) is a way of describing the rate at which null hypothesis testing produces errors. It has since become an essential tool for interpreting large datasets. In recent years, as datasets have become ever larger, and as the importance of 'big data' to scientific research has grown, the significance of the FDR has grown correspondingly.
The False Discovery Rate provides an analysis of the FDR's value as a tool, including why it should generally be preferred to the Bonferroni correction and other methods by which multiplicity can be accounted for. It offers a systematic overview of the FDR, its core claims, and its applications.
Readers of The False Discovery Rate will also find:
The False Discovery Rate is ideal for Statistics and Data Science courses, and short courses associated with conferences. It is also useful as supplementary reading in courses in other disciplines that require the statistical interpretation of 'big data'. The book will also be of great value to statisticians and researchers looking to learn more about the FDR.
STATISTICS IN PRACTICE
A series of practical books outlining the use of statistical techniques in a wide range of applications areas:
N. W. GALWEY is a Statistics Leader, Research Statistics, at GlaxoSmithKline Research and Development (Retired).
In the beginning was the significance threshold. By the early twentieth century, researchers with an awareness of random variation became concerned that the interesting results that they wished to report might have occurred by chance. Mathematicians worked to develop methods for quantifying this risk, and in 1926, R.A. Fisher wrote,
.it is convenient to draw the line at about the level at which we can say: 'Either there is something in the treatment, or a coincidence has occurred such as does not occur more than once in twenty trials'.
That is, he suggested a threshold of a?=?0.05, adding,
If one in twenty does not seem high enough odds, we may, if we prefer it, draw the line at one in fifty (the 2 per cent. point) or one in a hundred (the 1 per cent. point).
(Fisher, 1926, p. 504)
Fisher's suggestion was taken up by researchers, but for the next few decades they did not usually calculate the probability of obtaining, by coincidence, the observed result of each particular study. Such a calculation required a substantial amount of work by a trained mathematician. Instead, the researcher calculated a test statistic - z, t, F, ?2 or r (the correlation coefficient), depending on the design of the study and the question asked - and compared the value obtained with a published table of values corresponding to particular thresholds, typically a?=?0.05, 0.01 and 0.001. For example, suppose that a researcher analysing data from an experiment obtained the result t?=?-3.1, with 8 degrees of freedom (d.f.?=?8). If they were interested in large effects either positive or negative, they would consult a table of values of the t statistic for a two-sided test, and find that P(|T8|?>?2.306)?=?0.05 and P(|T8|?>?3.355)?=?0.01. Hence, they would conclude that their result was significant at the 5% (a?=?0.05) level, but not at the 1% (a?=?0.01) level. Though no probability had been calculated, such a conclusion could be reported in terms of a p-value - in this case, p?<?0.05.
By the late 1970s, many researchers had desktop or pocket calculators offering statistical functions, or even had access to programmable computers. This enabled them to present the actual p-value associated with their result, rather than comparing the result to pre-specified thresholds. In the present case, they would report p?=?0.015. However, the preoccupation with thresholds that had its origin in arithmetical convenience persisted, and a value of p?>?0.05 was (and is) typically presented as 'non-significant' ('NS'), whereas p?<?0.05 is 'significant' (often indicated by '*'); p?<?0.01 is 'highly significant' ('**'); and p?<?0.001 is 'very highly significant' ('***').
By this time, such significance tests had become the mainstay of statistical data analysis in the biological and social sciences - a status that they still retain. However, it was apparent from the outset that there are conceptual problems associated with such tests. Firstly, the test does not address precisely the question that the researcher most wants to answer. The researcher is not primarily interested in the probability of their dataset - in a sense its probability is irrelevant, as it is an event that has actually happened. What they really want to know is the probability of the hypothesis that the experiment was designed to test. This is the problem of 'inverse' or 'Bayesian' probability, the probability of things that are not - and cannot be - observed. Secondly, although the probability that a single experiment will give a significant result by coincidence is low, if more tests are conducted, the probability that at least one of them will do so increases.
Initially, these difficulties were dealt with by an informal understanding that if results were unlikely to be obtained by coincidence, then the probability that they were indeed produced by coincidence was low, and hence the hypothesis that this had occurred - the null hypothesis, H0 - could be rejected. It followed that among all the 'discoveries' announced by many researchers working over many years, it would not often turn out that the null hypothesis had, after all, been correct. As long as every study and every statistical analysis conducted required a considerable effort on the part of the researcher, there was some reason for confidence in this argument: researchers would not usually waste their time and other resources striving to detect effects that were unlikely to exist.
However, in the last decades of the twentieth century, technological developments changed the situation. There were several aspects to this expansion of the scope for statistical analysis, namely:
By the 1990s, the term 'big data' started to be used to refer to such developments. The management, manipulation and exploration of these huge datasets were characterised as a discipline called 'data science', distinct from classical statistics:
While the term data science is not new, the meanings and connotations have changed over time. The word first appeared in the '60s as an alternative name for statistics. In the late '90s, computer science professionals formalized the term. A proposed definition for data science saw it as a separate field with three aspects: data design, collection, and analysis. It still took another decade for the term to be used outside of academia. (Amazon Web Services, https://aws.amazon.com/what-is/data-science/, accessed 15 April 2024)
When thousands of statistical hypothesis tests could be performed with negligible effort either in the collection or the analysis of the data, the prospect that multiple testing would lead to significant results in cases where H0 was true - false positives - became effectively a certainty.
The problem of false-positive results is exacerbated if the multiple testing that has caused it is not apparent when the results are reported. This can occur, inadvertently or deliberately, due to several distinct mechanisms, for example, as follows:
Dateiformat: ePUBKopierschutz: Adobe-DRM (Digital Rights Management)
Systemvoraussetzungen:
Das Dateiformat ePUB ist sehr gut für Romane und Sachbücher geeignet – also für „fließenden” Text ohne komplexes Layout. Bei E-Readern oder Smartphones passt sich der Zeilen- und Seitenumbruch automatisch den kleinen Displays an. Mit Adobe-DRM wird hier ein „harter” Kopierschutz verwendet. Wenn die notwendigen Voraussetzungen nicht vorliegen, können Sie das E-Book leider nicht öffnen. Daher müssen Sie bereits vor dem Download Ihre Lese-Hardware vorbereiten.Bitte beachten Sie: Wir empfehlen Ihnen unbedingt nach Installation der Lese-Software diese mit Ihrer persönlichen Adobe-ID zu autorisieren!
Weitere Informationen finden Sie in unserer E-Book Hilfe.