Go to journal home page - Statistics & Probability Letters

Statistics & Probability Letters

About the journal, aims & scope.

Statistics & Probability Letters adopts a novel and highly innovative approach to the publication of research findings in statistics and probability . It features concise articles, rapid publication and broad coverage of the statistics and probability literature.

Statistics & Probability Letters is a refereed journal. Articles will be limited to six journal pages (13 double-space typed pages) including references and figures. Apart from the six-page limitation, originality, quality and clarity will be the criteria for choosing the material to be published in Statistics & Probability Letters . Every attempt will be made to provide the first review of a submitted manuscript within three months of submission .

The proliferation of literature and long publication delays have made it difficult for researchers and practitioners to keep up with new developments outside of, or even within, their specialization. The aim of Statistics & Probability Letters is to help to alleviate this problem. Concise communications (letters) allow readers to quickly and easily digest large amounts of material and to stay up-to-date with developments in all areas of statistics and probability.

The mainstream of Letters will focus on new statistical methods , theoretical results, and innovative applications of statistics and probability to other scientific disciplines. Key results and central ideas must be presented in a clear and concise manner. These results may be part of a larger study that the author will submit at a later time as a full length paper to SPL or to another journal. Theory and methodology may be published with proofs omitted, or only sketched, but only if sufficient support material is provided so that the findings can be verified. Empirical and computational results that are of significant value will be published. We also plan to publish applications and case studies that demonstrate a novel use of existing techniques or have interesting innovative ideas about data collection, modelling or inference.

Statistics & Probability Letters is …

Article publishing charge for open access

Co-Editors-in-Chief

Elton pei hsu.

Northwestern University Department of Mathematics, Evanston, Illinois, United States of America

Photo of Alessio Farcomeni, PhD

Alessio Farcomeni, PhD

University of Rome Tor Vergata Department of Economics and Finance, Roma, Italy

Latest published

Articles in press, most downloaded, most popular, penalized composite likelihood estimation for hidden markov models with unknown number of states, on absolute moment-based upper bounds for l-moments, unified specification tests in partially linear quantile regression models, convergence rates in the limit theorems for random sums of m -orthogonal random variables, input estimation from discrete workload observations in a lévy-driven storage system, block structure-based covariance tensor decomposition for group identification in matrix variables, on the moments of dividends and capital injections under a variant type of parisian ruin, consistent tests for semiparametric conditional independence, more from statistics & probability letters, announcements, call for papers, isi karl pearson prize for 2017, econometrics and statistics first issue is out., isi karl pearson prize for 2015, calls for papers, e-values and multiple testing, special issues and article collections, advances in statistical methods and applications for climate change and environment, the role of statistics in the era of big data, statistics in biological and medical sciences, silver jubilee issue dedicated to richard a. johnson on his 70th birthday.

Copyright © 2024 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • View all journals

Statistics articles from across Nature Portfolio

Statistics is the application of mathematical concepts to understanding and analysing large collections of data. A central tenet of statistics is to describe the variations in a data set or population using probability distributions. This analysis aids understanding of what underlies these variations and enables predictions of future changes.

Latest Research and Reviews

research paper about statistics and probability

Linear discriminant analysis

Linear discriminant analysis (LDA) is a versatile statistical method for reducing redundant and noisy information from an original sample to its essential features. In this Primer, Zhao et al. discuss LDA variants and their implementation settings as well as best practices for applying LDA to analyse data.

  • Shuping Zhao

research paper about statistics and probability

Effectiveness of non-pharmaceutical interventions for COVID-19 in USA

  • Weihao Wang

research paper about statistics and probability

Prediction of fresh herbage yield using data mining techniques with limited plant quality parameters

  • Şenol Çelik
  • Halit Tutar

research paper about statistics and probability

An identification method of LBL underwater positioning systematic error with optimal selection criterion

  • Jiongqi Wang
  • Xuanying Zhou

research paper about statistics and probability

Analyzing spatio-temporal dynamics of dissolved oxygen for the River Thames using superstatistical methods and machine learning

  • Takuya Boehringer
  • Christian Beck

research paper about statistics and probability

Passive earth pressure on vertical rigid walls with negative wall friction coupling statically admissible stress field and soft computing

  • Tram Bui-Ngoc

Advertisement

News and Comment

research paper about statistics and probability

Machine learning reveals the merging history of nearby galaxies

A probabilistic machine learning method trained on cosmological simulations is used to determine whether stars in 10,000 nearby galaxies formed internally or were accreted from other galaxies during merging events. The model predicts that only 20% of the stellar mass in present day galaxies is the result of past mergers.

research paper about statistics and probability

Efficient learning of many-body systems

The Hamiltonian describing a quantum many-body system can be learned using measurements in thermal equilibrium. Now, a learning algorithm applicable to many natural systems has been found that requires exponentially fewer measurements than existing methods.

research paper about statistics and probability

Fudging the volcano-plot without dredging the data

Selecting omic biomarkers using both their effect size and their differential status significance ( i.e. , selecting the “volcano-plot outer spray”) has long been equally biologically relevant and statistically troublesome. However, recent proposals are paving the way to resolving this dilemma.

  • Thomas Burger

research paper about statistics and probability

Disentangling truth from bias in naturally occurring data

A technique that leverages duplicate records in crowdsourcing data could help to mitigate the effects of biases in research and services that are dependent on government records.

  • Daniel T. O’Brien

research paper about statistics and probability

Sciama’s argument on life in a random universe and distinguishing apples from oranges

Dennis Sciama has argued that the existence of life depends on many quantities—the fundamental constants—so in a random universe life should be highly unlikely. However, without full knowledge of these constants, his argument implies a universe that could appear to be ‘intelligently designed’.

  • Zhi-Wei Wang
  • Samuel L. Braunstein

research paper about statistics and probability

A method for generating constrained surrogate power laws

A paper in Physical Review X presents a method for numerically generating data sequences that are as likely to be observed under a power law as a given observed dataset.

  • Zoe Budrikis

Quick links

  • Explore articles by subject
  • Guide to authors
  • Editorial policies

research paper about statistics and probability

Information

  • Author Services

Initiatives

You are accessing a machine-readable page. In order to be human-readable, please install an RSS reader.

All articles published by MDPI are made immediately available worldwide under an open access license. No special permission is required to reuse all or part of the article published by MDPI, including figures and tables. For articles published under an open access Creative Common CC BY license, any part of the article may be reused without permission provided that the original article is clearly cited. For more information, please refer to https://www.mdpi.com/openaccess .

Feature papers represent the most advanced research with significant potential for high impact in the field. A Feature Paper should be a substantial original Article that involves several techniques or approaches, provides an outlook for future research directions and describes possible research applications.

Feature papers are submitted upon individual invitation or recommendation by the scientific editors and must receive positive feedback from the reviewers.

Editor’s Choice articles are based on recommendations by the scientific editors of MDPI journals from around the world. Editors select a small number of articles recently published in the journal that they believe will be particularly interesting to readers, or important in the respective research area. The aim is to provide a snapshot of some of the most exciting work published in the various research areas of the journal.

Original Submission Date Received: .

  • Active Journals
  • Find a Journal
  • Journal Proposal
  • Proceedings Series
  • For Authors
  • For Reviewers
  • For Editors
  • For Librarians
  • For Publishers
  • For Societies
  • For Conference Organizers
  • Open Access Policy
  • Institutional Open Access Program
  • Special Issues Guidelines
  • Editorial Process
  • Research and Publication Ethics
  • Article Processing Charges
  • Testimonials
  • Preprints.org
  • SciProfiles
  • Encyclopedia

mathematics-logo

Journal Menu

  • Mathematics Home
  • Aims & Scope
  • Editorial Board
  • Reviewer Board
  • Topical Advisory Panel
  • Instructions for Authors
  • Special Issues
  • Sections & Collections
  • Article Processing Charge
  • Indexing & Archiving
  • Editor’s Choice Articles
  • Most Cited & Viewed
  • Journal Statistics
  • Journal History
  • Journal Awards
  • Society Collaborations
  • Conferences
  • Editorial Office

Journal Browser

  • arrow_forward_ios Forthcoming issue arrow_forward_ios Current issue
  • Vol. 12 (2024)
  • Vol. 11 (2023)
  • Vol. 10 (2022)
  • Vol. 9 (2021)
  • Vol. 8 (2020)
  • Vol. 7 (2019)
  • Vol. 6 (2018)
  • Vol. 5 (2017)
  • Vol. 4 (2016)
  • Vol. 3 (2015)
  • Vol. 2 (2014)
  • Vol. 1 (2013)

Find support for a specific problem in the support section of our website.

Please let us know what you think of our products and services.

Visit our dedicated information section to learn more about MDPI.

Probability, Statistics and Their Applications

  • Special Issue Editors

Special Issue Information

Benefits of publishing in a special issue.

  • Published Papers

A special issue of Mathematics (ISSN 2227-7390). This special issue belongs to the section " Probability and Statistics ".

Deadline for manuscript submissions: closed (31 December 2020) | Viewed by 101296

Share This Special Issue

Special issue editor.

research paper about statistics and probability

Dear Colleagues,

Statistics and probability are important domains in the scientific world, having many applications in various fields, such as engineering, reliability, medicine, biology, economics, physics, and not only, probability laws providing an estimated image of the world we live in.  This Special Volume deals targets some certain directions of the two domains as described below. 

Some applications of statistics are clustering of random variables based on simulated and real data or scan statistics, the latter being introduced in 1963 by Joseph Naus. In reliability theory, some important statistical tools are hazard rate and survival functions, order statistics, and stochastic orders. In physics, the concept of entropy is at its core, while special statistics were introduced and developed, such as statistical mechanics and Tsallis statistics.

~In economics, statistics, mathematics, and economics formed a particular domain called econometrics. ARMA models, linear regressions, income analysis, and stochastic processes are discussed and analyzed in the context of real economic processes. Other important tools are Lorenz curves and broken stick models.

~Theoretical results such as modeling of discretization of random variables and estimation of parameters of new and old statistical models are welcome, some important probability laws being heavy-tailed distributions. In recent years, many distributions along with their properties have been introduced in order to better fit the growing data available.

The purpose of this Special Issue is to provide a collection of articles that reflect the importance of statistics and probability in applied scientific domains. Papers providing theoretical methodologies and applications in statistics are welcome.

Prof. Vasile Preda Guest Editor

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website . Once you are registered, click here to go to the submission form . Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Mathematics is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2600 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

  • Applied and theoretical statistics
  • New probability distributions and estimation methods
  • Broken stick models
  • Lorenz curve
  • Scan statistics
  • Discretization of random variables
  • Clustering of random variables
  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here .

Published Papers (35 papers)

research paper about statistics and probability

Further Information

Mdpi initiatives, follow mdpi.

MDPI

Subscribe to receive issue release notifications and newsletters from MDPI journals

Statistics and Probability Have Always Been Value-Laden: An Historical Ontology of Quantitative Research Methods

  • Original Paper
  • Published: 27 May 2019
  • Volume 167 , pages 1–18, ( 2020 )

Cite this article

research paper about statistics and probability

  • Michael J. Zyphur 1 &
  • Dean C. Pierides 2  

5217 Accesses

17 Citations

12 Altmetric

Explore all metrics

Quantitative researchers often discuss research ethics as if specific ethical problems can be reduced to abstract normative logics (e.g., virtue ethics, utilitarianism, deontology). Such approaches overlook how values are embedded in every aspect of quantitative methods, including ‘observations,’ ‘facts,’ and notions of ‘objectivity.’ We describe how quantitative research practices, concepts, discourses, and their objects/subjects of study have always been value-laden, from the invention of statistics and probability in the 1600s to their subsequent adoption as a logic made to appear as if it exists prior to, and separate from, ethics and values. This logic, which was embraced in the Academy of Management from the 1960s, casts management researchers as ethical agents who ought to know about a reality conceptualized as naturally existing in the image of statistics and probability (replete with ‘constructs’), while overlooking that S&P logic and practices, which researchers made for themselves, have an appreciable role in making the world appear this way. We introduce a different way to conceptualize reality and ethics, wherein the process of scientific inquiry itself requires an examination of its own practices and commitments. Instead of resorting to decontextualized notions of ‘rigor’ and its ‘best practices,’ quantitative researchers can adopt more purposeful ways to reason about the ethics and relevance of their methods and their science. We end by considering implications for addressing ‘post truth’ and ‘alternative facts’ problems as collective concerns, wherein it is actually the pluralistic nature of description that makes defending a collectively valuable version of reality so important and urgent.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save.

  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime

Price includes VAT (Russian Federation)

Instant access to the full article PDF.

Rent this article via DeepDyve

Institutional subscriptions

Similar content being viewed by others

research paper about statistics and probability

Data Analysis Techniques for Quantitative Study

research paper about statistics and probability

Case Studies and Statistics in Causal Analysis: The Role of Bayesian Narratives

research paper about statistics and probability

Statistics in the Service of Science: Don’t Let the Tail Wag the Dog

Explore related subjects.

  • Artificial Intelligence
  • Medical Ethics

Abbott, A. (1991). The order of professionalization: An empirical analysis. Work and Occupations, 18, 355–384.

Google Scholar  

Abbott, A. (1998). The causal devolution. Sociological Methods and Research, 27, 148–181.

Adams, W. J. (2009). The life and times of the central limit theorem . Providence, RI: American Mathematical Society.

Aguinis, H., Werner, S., Abbott, J. L., Angert, C., Park, J. H., & Kohlhausen, D. (2010). Customer-centric science: Reporting significant research results with rigor, relevance, and practical impact in mind. Organizational Research Methods, 13, 515–539.

Alanen, L. (2003). Descartes’s concept of mind . Cambridge, MA: Harvard University Press.

Aldrich, J. (1997). Fisher and the making of maximum likelihood 1912-1922. Statistical Science, 12, 133–220.

Aldrich, J. (2008). R. A. Fisher on Bayes and Bayes’ theorem. Bayesian Analysis, 3, 161–170.

Alonso, W., & Starr, P. (Eds.). (1987). The politics of large numbers . New York: Russell Sage.

American Psychiatric Association. (2013). Diagnostic and statistical manual of mental disorders (5th ed.). Arlington, VA: American Psychiatric Publishing.

Ariew, A. (2007). Under the influence of Malthus’s law of population growth: Darwin eschews the statistical techniques of Adolphe Quetelet. Studies in the History and Philosophy of Biology and Biomedicine, 38, 1–19.

Arnauld, A., & Nicole, P. (1662/1996). Logic or the art of thinking . Cambridge, UK: Cambridge University Press.

Augier, M., & March, J. G. (2011). The roots, rituals, and rhetorics of change: North American business schools after the second world war . Stanford, CA: Stanford University Press.

Bayes, T. P. (1763). An essay towards solving a problem in the doctrine of chances. Philosophical Transactions of the Royal Society of London, 53, 370–418.

Bernoulli, J. (1713/2006). The art of conjecturing (together with letter to a friend on sets in court tennis) (E. D. Sylla, Trans. & Ed.). Baltimore, MD: Johns Hopkins University Press.

Bernthal, W. F. (1960). Integrating the behavioral sciences and management. The Journal of the Academy of Management, 3, 161–166.

Biagioli, M. (2002). From book censorship to academic peer review. Emergences: Journal for the Study of Media and Composite Cultures, 12, 11–45.

Biagioli, M. (2006). Galileo’s instruments of credit: Telescopes, images, secrecy . Chicago, IL: University of Chicago Press.

Bornemann, A. (1961). The development of the teaching of management in the school of business. The Journal of the Academy of Management, 4, 129–136.

Bromhead, H. (2009). The reign of truth and faith: Epistemic expressions in the 16th and 17th century English . Berlin: Mouton de Gruyter.

Byrne, E. F. (1968). Probability and opinion: A study of the medieval presuppositions of post-medieval theories of probability . The Hague: Matinus Nijhoff.

Casler, C., & du Gay, P. (2019). Stances, paradigms, personae. Studi di Sociologia, 1, 69–80.

Caton, H. (1973). The origin of subjectivity . New Haven, CT: Yale University Press.

Cohen, I. B. (1987). Scientific revolutions, revolutions in science, and a probabilistic revolution 1800-1930. In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 23–44). Cambridge, MA: MIT Press.

Cooke, B., & Alcadipani, R. (2015). Toward a global history of management education: The case of the Ford Foundation and the São Paulo School of Business Administration, Brazil. Academy of Management Learning & Education, 14, 482–499.

Cortina, J. (2019). On the whys and hows of quantitative research. Journal of Business Ethics . https://doi.org/10.1007/s10551-019-04195-8 .

Article   Google Scholar  

Dale, A. I. (1999). A history of inverse probability: From Thomas Bayes to Karl Pearson (2nd ed.). New York: Springer.

Danziger, K. (1990). Constructing the subject: Historical origins of psychological research . Cambridge, UK: Cambridge University Press.

Daston, L. (1987a). The domestication of risk: Mathematical probability and insurance 1650-1830. In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 238–260). Cambridge, MA: MIT Press.

Daston, L. (1987b). Rational individuals versus laws of society: From probability to statistics. In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 295–304). Cambridge, MA: MIT Press.

Daston, L. (1992). Objectivity and the escape from perspective. Social Studies of Science, 22, 597–618.

Daston, L. (1994). How probability came to be objective and subjective. Historia Mathematica, 21, 330–344.

Daston, L. (1995). Classical probability in the enlightenment . Princeton, NJ: Princeton University Press.

Daston, L. (2005). Scientific error and the ethos of belief. Social Research, 72, 1–28.

Daston, L., & Galison, P. (2007). Objectivity . Brooklyn, NY: Zone Books.

Daston, L., & Vidal, F. (Eds.). (2004). The moral authority of nature . Chicago, IL: University of Chicago Press.

De Morgan, A. (1838/2010). An essay on probabilities: And their application to life contingencies and insurance offices . Charleston, SC: BiblioLife, LLC.

Dear, P. (2001). Revolutionizing the sciences: European knowledge and its ambitions, 1500-1700 . London: Palgrave.

Desrosières, A. (1998). The politics of large numbers: A history of statistical reasoning , C. Naish (trans.). Cambridge, UK: Harvard University Press.

Dewey, J. (1929). The quest for certainty . New York: Minton, Balch, & Co.

Dewey, J. (1938). Logic: The theory of inquiry . New York: Henry Holt and Company.

Douglas, H. E. (2009). Science, policy, and the value-free ideal . Pittsburgh, PA: University of Pittsburgh Press.

Duncan, O. D. (1984). Notes on social measurement: Historical and critical . New York: Russell Sage.

Edwards, J. R. (2010). Reconsidering theoretical progress in organizational and management research. Organizational Research Methods, 13, 615–619.

Edwards, J. R. (2019). The peaceful coexistence of ethics and quantitative research. Journal of Business Ethics . https://doi.org/10.1007/s10551-019-04197-6 .

Ezzamel, M., & Willmott, H. (2014). Registering ‘the ethical’ in organization theory formation: Towards the disclosure of an ‘invisible force’. Organization Studies, 35, 1–27.

Fischer, H. (2010). A history of the central limit theorem: From classical to modern probability theory . New York: Springer.

Fisher, R. A. (1915). The frequency distribution of the values of the correlation coefficient in samples from an indefinitely large population. Biometrika, 10, 507–521.

Fisher, R. A. (1921). On the “probable error” of a correlation coefficient deduced from a small sample. Metron, 1, 3–32.

Fisher, R. A. (1922a). On the mathematical foundations of theoretical statistics. Philosophical Transactions of the Royal Society of London, A, 222, 309–368.

Fisher, R. A. (1922b). The goodness of fit of regression formulae and the distribution of regression coefficients. Journal of the Royal Statistical Society, 85, 597–612.

Fisher, R. A. (1924). The distribution of the partial correlation coefficient. Metron, 3, 329–332.

Fisher, R. A. (1925a). Applications of “Student’s” distribution. Metron, 5, 90–104.

Fisher, R. A. (1925b). Statistical methods for research workers . Edinburgh: Oliver & Boyd Ltd.

Fisher, R. A. (1925c). Theory of statistical estimation. Proceedings of the Cambridge Philosophical Society, 22, 700–725.

Fisher, R. A. (1928). The general sampling distribution of the multiple correlation coefficient. Proceedings of the Royal Society, A, 121, 654–673.

Fisher, R. A. (1929). Moments and product moment of sampling distributions. Proceedings of the London Mathematical Society, Series, 2 (30), 199–238.

Fisher, R. A. (1935a). The design of experiments . Edinburgh: Oliver & Boyd Ltd.

Fisher, R. A. (1935b). The logic of inductive inference. Journal of the Royal Statistical Society, 98, 39–54.

Fisher, R. A. (1955). Statistical methods and scientific induction. Journal of the Royal Statistical Society: Series B, 17, 69–78.

Fisher, R. A. (1956). Statistical methods and scientific inference . Edinburgh: Oliver & Boyd.

Fisher, R. A., & Yates, F. (1943). Statistical tables for biological, agricultural and medical research . Edinburgh, UK: Oliver & Boyd Ltd.

Foucault, M. (1970). The order of things: An archaeology of the human sciences . New York: Random House.

Foucault, M. (1980). Power/knowledge: Selected interviews and other writings . New York: Random House.

Foucault, M. (2003). Abnormal: Lectures at the college de France 1974-1975 . New York: Graham Burchell.

Foucault, M. (2008). The birth of biopolitics Lectures at the college de France 1978-1979 . New York: Graham Burchell.

Freedman, D. A. (2005). Linear statistical models for causation: A critical review. In B. S. Everitt & D. C. Howell (Eds.), Encyclopedia of statistics in behavioral science (Vol. 2, pp. 1061–1073). Chichester, UK: Wiley.

Galavotti, M. C. (2005). Philosophical introduction to probability . Palo Alto, CA: CSLI.

Galileo, G. (2008). The essential Galileo (M. A. Finocchiaro, Ed.). Indianapolis, IN: Hackett Publishing.

Galton, G. (1869/2010). Hereditary genius: An inquiry into its laws and consequences. Memphis, TN: General Books.

Galton, F. (1889/2010). Natural inheritance . Charleston, SC: Nabu Press.

Garber, D. (1992). Descartes’ physics. In J. Cottingham (Ed.), The Cambridge companion to Descartes . Cambridge, UK: Cambridge University Press.

Gauss, C. F. (1809/2010). Theory of the motion of the heavenly bodies moving about the sun in conic sections (C. H. Davis, Trans.). Charleston, SC: Nabu Press.

Gigerenzer, G. (1998). We need statistical thinking, not statistical rules. Behavioral and Brain Sciences, 21, 199–200.

Gigerenzer, G. (2004). Mindless statistics. The Journal of Socio-Economics, 33, 587–606.

Gigerenzer, G., & Marewski, J. N. (2015). Surrogate science: The idol of a universal method for scientific inference. Journal of Management, 41, 421–440.

Gigerenzer, G., & Murray, D. J. (1987). Cognition as intuitive statistics . Hillsdale, MI: Erlbaum.

Gigerenzer, G., Swijtink, Z., Porter, T., Daston, L., Beatty, J., & Krüger, L. (1989). The empire of chance. How probability changed science and everyday life . Cambridge, UK: Cambridge University Press.

Goldman, L. (1983). The origins of British ‘social science’: Political economy, natural science and statistics, 1830-1835. The Historical Journal, 26, 587–616.

Golembiewski, R. T. (1961). Management science and group behavior: Work-unit cohesiveness. Journal of the Academy of Management, 4, 87–99.

Gordon, R. A., & Howell, J. E. (1959). Higher education for business . New York: Columbia University Press.

Greenwood, M. (2016). Approving or improving research ethics in management journals. Journal of Business Ethics, 137, 1–14.

Hacking, I. (1965). Logic of statistical inference . Cambridge, UK: Cambridge University Press.

Hacking, I. (1975). Why does language matter to philosophy? . Cambridge, UK: Cambridge University Press.

Hacking, I. (1978). Hume’s species of probability. Philosophical Studies, 33, 21–37.

Hacking, I. (1986). The archaeology of Foucault. In D. C. Hoy (Ed.), Foucault: A critical reader (pp. 27–40). Oxford, UK: Blackwell.

Hacking, I. (1987). Was there a probabilistic revolution 1800-1930? In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 45–55). Cambridge, MA: MIT Press.

Hacking, I. (1990). The taming of chance . Cambridge, UK: Cambridge University Press.

Hacking, I. (1991). How should we do the history of statistics? In G. Burchell, C. Gordon, & P. Miller (Eds.), The Foucault effect (pp. 181–195). Chicago, IL: Chicago University Press.

Hacking, I. (1992). Statistical language, statistical truth and statistical reason: The self-authentification of a style of scientific reasoning. In E. McMullin (Ed.), The social dimensions of science (Vol. 3, pp. 130–157). Notre Dame, IN: University of Notre Dame Press.

Hacking, I. (1999). The social construction of what? . Cambridge, UK: Harvard University Press.

Hacking, I. (2002). Historical ontology . Cambridge, UK: Harvard University Press.

Hacking, I. (2006). The emergence of probability: A philosophical study of early ideas about probability, induction and statistical inference (2nd ed.). Cambridge, UK: Cambridge University Press.

Hacking, I. (2009). Scientific reasoning . Taipei: NTU Press.

Hald, A. (2007). A history of parametric statistical inference from Bernoulli to Fisher, 1713-1935 . New York: Springer.

Halff, J. F. (1960). Applying the scientific method to the study of management. The Journal of the Academy of Management, 3, 193–196.

Hardy, C., Phillips, N., & Clegg, S. R. (2001). Reflexivity in organization and management theory: A study of the production of the research ‘subject’. Human Relations, 54, 531–560.

Hennig, B. (2007). Cartesian conscientia. British Journal for the History of Philosophy, 15, 455–484.

Henry, J. (2004). Metaphysics and the origins of modern science: Descartes and the importance of laws of nature. Early Science and Medicine, 9, 73–114.

Honig, B., Lampel, J., Siegel, D., & Drnevich, P. (2017). Special section on ethics in management research: Norms, identity, and community in the 21st century. Academy of Management Learning & Education, 16, 84–93.

Hopwood, N., Schaffer, S., & Secord, J. (2010). Seriality and scientific objects in the nineteenth century. History of Science, 48, 251–284.

House, R. J. (1962). An experiment in the use of management training standards. Journal of the Academy of Management, 5, 76–81.

House, R. J. (1975). The quest for relevance in management education: Some second thoughts and undesired consequences. Academy of Management Journal, 18, 323–333.

Howie, D. (2002). Interpreting probability: Controversies and developments in the early twentieth century . Cambridge, UK: Cambridge University Press.

Hume, D. (1739). A treatise of human nature: Being an attempt to introduce the experimental method of reasoning into moral subjects (Vol. 1). London: Jon Noon.

Huygens, C. (1657/2010). De ratiociniis in aleae ludo. or, the value of all chances in games of fortune; … mathematically demonstrated . Farmington Hills, MI: Gale ECCO.

Kamlah, A. (1987). The decline of the Laplacian theory of probability: A study of Stupf, von Kries, and Meinong. In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 91–116). Cambridge, MA: MIT Press.

Khurana, R. (2007). From hired aims to hired hands: The social transformation of American business schools and the unfulfilled promise of management as a profession . Princeton, NJ: Princeton University Press.

Köhler, T., Landis, R. S., & Cortina, J. M. (2017). From the editors: Establishing methodological rigor in quantitative management learning and education research: The role of design, statistical methods, and reporting standards. Academy of Management Learning & Education, 16, 173–192.

Krüger, L. (1987). The slow rise of probabilism: Philosophical arguments in the nineteenth century. In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 59–89). Cambridge, MA: MIT Press.

Krüger, L., Gigerenzer, G., & Morgan, M. S. (Eds.). (1987). The probabilistic revolution: Ideas in the sciences (Vol. 2). Cambridge, MA: MIT Press.

Kuhn, T. S. (2012). The structure of scientific revolutions (5th ed.). Chicago, IL: University of Chicago Press.

Lagrange, J. L. (1770–1773/2009). Memoir on the utility of the method of taking the mean among the results of several observations in which one examines the advantage of this method by the calculus of probabilities, and where on selves different problems related to this material (R. J. Pulskamp, Trans.) http://cerebro.xu.edu/math/Sources/Lagrange/on%20mean.pdf .

Laplace, P.-S. (1774/1986). Memoir on the probability of the cause of events (S. Stigler, Trans.). Statistical Science , 1 , 364–378.

Laplace, P.-S. (1809a/2011). Mémoire sur les approximations des formules qui sont fonctions de très - grands nombres, et sur leur application aux probabilities (R. J. Pulskamp, Trans.). http://www.cs.xu.edu/math/Sources/Laplace/approximations%20of%20formulas%201809.pdf .

Laplace, P.-S. (1809b/2010). Supplement au mémoire sur les approximations des formules qui sont fonctions de très - grands nombres (R. J. Pulskamp, Trans.). http://www.cs.xu.edu/math/Sources/Laplace/supplement%201809.pdf .

Laplace, P.-S. (1825/1995). Philosophical essay on probability (A. I. Dale, Trans.). New York: Springer.

Lécuyer, B.-P. (1987). Probability in vital and social statistics: Quetelet, Farr, and the Bertillons. In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 317–335). Cambridge, MA: MIT Press.

Leibniz, G. (1678/2004). Leibniz on estimating the uncertain: An English translation of De incerti aestimatione with commentary (W. D. C. de Leo and J. Cussens, Trans.). The Leibniz Review, 14 , 31–53.

MacKenzie, D. (1981). Statistics in Britain, 1865-1930: The social construction of scientific knowledge . Edinburgh, UK: Edinburgh University Press.

Mandelbaum, M. (1964). Philosophy, science, and sense perception . Baltimore, MD: Johns Hopkins Press.

Martela, F. (2015). Fallible inquiry with ethical ends-in-view: A pragmatist philosophy of science for organizational research. Organization Studies, 36, 537–563.

McFarland, D. E. (1960). The emerging revolution in management education. The Journal of the Academy of Management, 3, 7–15.

McGrayne, S. (2011). The theory that would not die . New Haven, CT: Yale University Press.

Melton, A. W. (1962). Editorial. Journal of Experimental Psychology, 64, 553–557.

Metz, K. H. (1987). Paupers and numbers: The statistical argument for social reform in Britain during the period of industrialization. In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 337–350). Cambridge, MA: MIT Press.

Miner, J. B. (1963). Psychology and the business school curriculum. Academy of Management Journal, 46, 284–289.

Moore, D. G. (1960). Behavioral science and business education. The Journal of the Academy of Management, 3, 187–191.

Neyman, J. (1934). On the two different aspects of the representative method: The method of stratified random sampling and the method of purposive selection (with discussion). Journal of the Royal Statistical Society, B, 97, 558–606.

Neyman, J. (1937). Outline of a theory of statistical estimation based on the classical theory of probability. Philosophical Transactions of the Royal Society of London, A, 236, 333–380.

Neyman, J. (1950). First course in probability and statistics . New York: Henry Holt.

Neyman, J. (1956). Note on an article by Sir Ronald Fisher. Journal of the Royal Statistical Society: Series B, 18, 288–294.

Neyman, J., & Pearson, E. S. (1933). On the problem of the most efficient tests of statistical hypotheses. Philosophical Transactions of the Royal Society, A, 231, 289–337.

Osler, M. J. (1970). John Locke and the changing ideal of scientific knowledge. Journal of the History of Ideas, 31, 3–16.

Panter, A. T., & Sterba, S. K. (Eds.). (2011). Handbook of ethics in quantitative methodology . New York: Taylor & Francis.

Pascal, B. (1653/1952). Treatise on the arithmetical triangle. In R. M. Hutchins (Ed.), Great books of the Western world: Pascal (Vol. 33, pp. 447–473). Chicago, IL: University of Chicago Press.

Pascal, B. (1654/1952). Correspondence with Fermat on the theory of probabilities. In R. M. Hutchins (Ed.), Great books of the Western world: Pascal (Vol. 33, pp. 474–487). Chicago, IL: University of Chicago Press.

Pearson, E. S. (1978). The history of statistics in the 17th & 18th centuries . London: Charles Griffin & Co.

Petit, V. (2013). An object called population. In V. Petit (Ed.), Counting populations, understanding societies (pp. 53–87). Dordrecht: Springer.

Pierson, F. C. (1959). The education of American businessmen: A study of university-college programmes in business administration . New York: McGraw-Hill.

Poovey, M. (1995). Making a social body: British cultural formation, 1830-1864 . Chicago, IL: Chicago University Press.

Poovey, M. (1998). A history of the modern fact: Problems of knowledge in the sciences of wealth and society . Chicago, IL: Chicago University Press.

Popper, K. (1959/2002). The logic of scientific discovery . London: Routledge.

Porter, T. M. (1986). The rise of statistical thinking, 1820-1900 . Princeton, NJ: Princeton University Press.

Porter, T. M. (1994). From Quetelet to Maxwell: Social statistics and the original of statistical physics. In I. B. Cohen (Ed.), The natural sciences and the social sciences (Vol. 150, pp. 345–362). New York: Springer.

Porter, T. M. (1995). Trust in numbers: The pursuit of objectivity in science and public life . Princeton, NJ: Princeton University Press.

Powell, T. (2019). Can quantitative research solve social problems? Pragmatism and the ethics of social research. Journal of Business Ethics . https://doi.org/10.1007/s10551-019-04196-7 .

Quetelet, L. A. J. (1842/2010). A treatise on man and the development of his faculties (R. Knox, Trans., T. Smibert, Ed.). Charleston, SC: Nabu Press.

Rorty, R. (1979). Philosophy and the mirror of nature . Princeton, NJ: Princeton University Press.

Ross, D. (1991). The origins of American social science . Cambridge, UK: Cambridge University Press.

Rowlinson, M., Hassard, J., & Decker, S. (2014). Research strategies for organizational history: A dialogue between historical theory and organization theory. Academy of Management Review, 39, 250–274.

Rucci, A. J., & Tweney, R. D. (1980). Analysis of variance and the “second discipline” of scientific psychology: A historical account. Psychological Bulletin, 87, 166–184.

Rynes, S. L., Bartunek, J. M., & Daft, R. L. (2001). Across the great divide: Knowledge creation and transfer between practitioners and academics. Academy of Management Journal, 44, 340–355.

Schlossman, S., & Sedlak, M. (1985). The age of autonomy in American management education. Selections, 1, 15–26.

Schlossman, S. L., Sedlak, M. W., & Wechsler, H. S. (1987). The ‘new look’: The Ford Foundation and the revolution in business education . Los Angeles, CA: Graduate Management Admissions Council.

Schneider, I. (1987). Laplace and thereafter: The status of probability calculus in the nineteenth century. In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 191–214). Cambridge, MA: MIT Press.

Schouls, P. A. (1979). The imposition of method: A study of Descartes and Locke . Oxford, UK: Oxford University Press.

Schouls, P. A. (2000). Descartes and the possibility of science . Ithaca, NY: Cornell University Press.

Schwab, A., & Starbuck, W. H. (2017). A call for openness in research reporting: How to turn covert practices into helpful tools. Academy of Management Learning & Education, 16 (1), 125–141.

Shapin, S. (1984). Pump and circumstance: Robert Boyle’s literary technology. Social Studies of Science, 14, 481–520.

Shapin, S. (1994). A social history of truth: Civility and science in seventeenth-century England . Chicago, IL: University of Chicago Press.

Shapin, S. (2008). The scientific life: A moral history of a late modern vocation . Chicago, IL: University of Chicago Press.

Shapin, S. (2010). Never pure: Historical studies of science as if it was produced by people with bodies, situated in time, space, culture, and society, and struggling for credibility and authority . Baltimore, MD: Johns Hopkins University Press.

Shapin, S., & Schaffer, S. (1985). Leviathan and the air-pump: Hobbes, Boyle, and the experimental life . Princeton, NJ: Princeton University Press.

Shapiro, B. J. (1983). Probability and certainty in seventeenth-century England: A study of the relationships between natural science, religion, history, law, and literature . Princeton, NJ: Princeton University Press.

Shull, F. (1962). The nature and contribution of administrative models and organizational research. The Journal of the Academy of Management, 5, 124–138.

Sterling, R. D. (1959). Publication decisions and their possible effects on inferences drawn from tests of significance—Or vice versa. Journal of the American Statistical Association, 54, 30–34.

Stigler, S. M. (1986). The history of statistics: The measurement of uncertainty before 1900 . Cambridge, UK: Harvard University Press.

Stigler, S. M. (1999). Statistics on the table: The history of statistical concepts and methods . Cambridge, UK: Harvard University Press.

Strong, J. V. (1978). John Stuart Mill, John Herschel, and the ‘probability of causes’. PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association, 1, 31–41.

Swijtink, Z. G. (1987). The objectification of observation: Measurement and statistical methods in the nineteenth century. In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 261–292). Cambridge, MA: MIT Press.

Tadajewski, M. (2009). The politics of the behavioral revolution. Organization, 16, 733–754.

Towle, J. W. (1960). Opportunities ahead for the Academy of Management. The Journal of the Academy of Management, 3, 147–154.

Truesdell, C. (1984). An idiot’s fugitive essays on science: Methods, criticism, training, circumstances . New York: Springer.

Venn, J. (1866/2006). The logic of chance . Mineola: Dover Publications.

Wald, A. (1945). Sequential tests of statistical hypotheses. The Annals of Mathematical Statistics, 16, 117–186.

Wald, A. (1950). Statistical decision functions . New York: Wiley.

Weatherbee, T. G. (2012). Caution! This historiography makes wide turns: Historic turns and breaks in management and organization studies. Management & Organizational History, 7, 203–218.

Wicks, A. C., & Freeman, R. E. (1998). Organizational studies and the new pragmatism: Positivism, anti-positivism, and the search for ethics. Organization Science, 9, 123–140.

Williams, B. (2005). Descartes: The project of pure inquiry . New York: Routledge.

Wise, M. N. (1987). How do sums count? One the cultural origins of statistical causality. In L. Krüger, L. J. Daston, & M. Heidelberger (Eds.), The probabilistic revolution (Vol. 1, pp. 395–425). Cambridge, MA: MIT Press.

Woolf, S. (1989). Statistics and the modern state. Comparative Studies in Society and History, 31, 588–604.

Yule, G. U. (1897). One the theory of correlation. Journal of the Royal Statistical Society, 60, 812–854.

Yule, G. U. (1899). An investigation into the causes of changes in pauperism in England, chiefly during the last two intercensal decades, I . Journal of the Royal Statistical Society, 62, 249–295.

Zabell, S. (1989). R. A. Fisher on the history of inverse probability. Statistical Science, 4, 247–263.

Zabell, S. (1992). R. A. Fisher and fiducial argument. Statistical Science, 7, 369–387.

Zyphur, M. J., & Pierides, D. C. (2017). Is quantitative research ethical? Tools for ethically practicing, evaluating, and using quantitative research. Journal of Business Ethics, 143, 1–16.

Zyphur, M. J., & Pierides, D. C. (2019). Making quantitative research work: From positivist dogma to actual social scientific inquiry. Journal of Business Ethics . https://doi.org/10.1007/s10551-019-04189-6 .

Download references

Acknowledgements

For reviewing previous versions of this manuscript the authors thank Adam Barsky, Andrew Gelman, Andrew Van de Ven, Barbara Lawrence, Fred Oswald, Graham Sewell, Karen Jehn, Kristopher Preacher, Maria Carla Galavotti, Ray Zammuto, Tom Lee, Zhen Zhang, and Dan Woodman. For help with historical insights, the authors thank James March, Steven Schlossman, William Starbuck, and the many independent presses that have made it possible to investigate the roots of our quantitative practices. We would also like to thank the anonymous reviewers who have read and responded to previous versions of this manuscript.

This research was supported by Australian Research Council’s Future Fellowship scheme (project FT140100629).

Author information

Authors and affiliations.

Department of Management and Marketing, University of Melbourne, Parkville, VIC, 3010, Australia

Michael J. Zyphur

University of Stirling, Stirling, FK9 4LA, UK

Dean C. Pierides

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Michael J. Zyphur .

Ethics declarations

Conflict of interest.

The authors declare no conflicts of interest.

Ethical Approval

This article does not contain any studies with human participants or animals performed by any of the authors.

Additional information

Publisher's note.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Zyphur, M.J., Pierides, D.C. Statistics and Probability Have Always Been Value-Laden: An Historical Ontology of Quantitative Research Methods. J Bus Ethics 167 , 1–18 (2020). https://doi.org/10.1007/s10551-019-04187-8

Download citation

Received : 18 April 2018

Accepted : 17 May 2019

Published : 27 May 2019

Issue Date : November 2020

DOI : https://doi.org/10.1007/s10551-019-04187-8

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Quantitative research methods
  • Research ethics
  • Historical ontology
  • Statistics and probability
  • Best practices
  • Questionable research practices

Advertisement

  • Find a journal
  • Publish with us
  • Track your research

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

The PMC website is updating on October 15, 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • J Thorac Dis
  • v.7(3); 2015 Mar

A practical overview on probability distributions

Andrea viti, alberto terzi, luca bertolaccini.

Aim of this paper is a general definition of probability, of its main mathematical features and the features it presents under particular circumstances. The behavior of probability is linked to the features of the phenomenon we would predict. This link can be defined probability distribution. Given the characteristics of phenomena (that we can also define variables), there are defined probability distribution. For categorical (or discrete) variables, the probability can be described by a binomial or Poisson distribution in the majority of cases. For continuous variables, the probability can be described by the most important distribution in statistics, the normal distribution. Distributions of probability are briefly described together with some examples for their possible application.

A short definition of probability

We can define the probability of a given event by evaluating, in previous observations, the incidence of the same event under circumstances that are as similar as possible to the circumstances we are observing [this is the frequentistic definition of probability, and is based on the relative frequency of an observed event, observed in previous circumstances ( 1 )]. In other words, probability describes the possibility of an event to occur given a series of circumstances (or under a series of pre-event factors). It is a form of inference, a way to predict what may happen, based on what happened before under the same (never exactly the same) circumstances. Probability can vary from 0 (our expected event was never observed, and should never happen) to 1 (or 100%, the event is almost sure). It is described by the following formula: if X = probability of a given x event ( Eq. [1] ):

This is one of the three axioms of probability, as described by Kolmogorov ( 2 ):

  • If under some circumstances, a given number of events ( E ) could verify ( E 1 , E 2 , E 3 , …, E n ), the probability ( P ) of any E is always more than zero;
  • The sum of the probabilities of E =  P ( E 1 ) +  P ( E 2 ) +  ⋯  +  P ( E n ) is 100%;
  • If E 1 and E 3 are two possible events, the probability that one or the other could happen P ( E 1 or E 3 ) is equal to the sum of the probability of E 1 and the probability of E 3 ( Eq. [2] ): P ( E 1   o r   E 2 ) =  P ( E 1 ) +  P ( E 3 ) [2]

Probability could be described by a formula, a graph, in which each event is linked to its probability. This kind of description of probability is called probability distribution.

Binomial distribution

A classic example of probability distribution is the binomial distribution. It is the representation of the probability when only two events may happen, that are mutually exclusive. The typical example is when you toss a coin. You can only have two results. In this case, the probability is 50% for both events. However, binomial distribution may describe also two events that are mutually exclusive but are not equally possible (for instance that a newborn baby will be left-handed or right-handed). The probability that x individuals present a given characteristic, p , that is mutually exclusive of another one, called q , depends on the possible number of combinations of x individuals within the population, called C. If my population is composed of five5 individuals, that can be p or q , I have ten possible combinations of, for instance, three individuals with p is ( Eq. [3] ):

pppqq , pqqpp , ppqpq , ppqqp, pqpqpq, qpppq, qpqpp, qppqp, qqppp

Then p 3 q 2 will be multiplied for the number of combinations (ten times).

If, in experimental population, I had a big number of individuals ( n ), the number of combinations of x individuals within the population will be ( Eq. [4] ):

Therefore, the probability that a group of x individuals within the population of n individuals presents the characteristic p , that excludes q , will be described by the following formula ( Eq. [5] ):

that describes the binomial distribution. It follows the Kolmogorow’s rules ( Eq. [6] ):

In a given population, 30% of the people are left-handed. If we select ten individuals from this population, what is the probability that four out of ten individuals are left handed?

We can apply the binomial distribution, since we suppose that a person may be either left-handed or right-handed.

Se we can use our formula ( Eq. [7] ):

Poisson distribution

Another important distribution of probability is the Poisson distribution. It is useful to describe the probability that a given event can happen within a given period (for instance, how many thoracic traumas could need the involvement of the thoracic surgeon in a day, or a week, etc.). The events that may be described by this distribution have the following characteristics:

  • The events are independent from one another;
  • Within a given interval the event may present from 0 to infinite times;
  • The probability of an event to happen increases when the period of observation is longer.

To predict the probability, I must know how the events behave (this data comes from previous, or historical, observations of the same event before the time I am trying to perform my analysis). This parameter, that is a mean of the events in a given interval, as derived from previous observations, is called λ .

The Poisson distribution follows the following formula ( Eq. [8] ):

where the number e is an important mathematical constant that is the base of the natural logarithm. It is approximately equal to 2.71828.

For example, the distribution of major thoracic traumas needing intensive care unit (ICU) recovery during a month in the last three years in a Third Level Trauma Center follows a Poisson distribution, were λ =2.75. In a future period of one month, what is the probability to have three patients with major thoracic trauma in ICU? ( Eq. [9] ):

Therefore, the probability is 22.1%.

The binomial distribution refers only to discrete variables (that present a limited number of values within a given interval). However, in nature, many variables may present an infinite distribution of values, within a given interval. These are called continuous variables ( 3 ).

Distributions of continuous variables

An example of continuous variable is the systolic blood pressure. Within a given cohort of systolic blood pressure can be presented as in Figure 1 . Each single histogram length represents an interval of the measure of interest between two intervals on the x -axis, while the histogram height represents the number of measured values within the interval. When the number of observation becomes very large (tends to infinite) and the length of the histogram becomes narrower (tends to 0), the above representation becomes more similar to a curved line ( Figure 2 ). This curve describes the distribution of probability, f (density of probability) for any given value of x , the continuous variable. The area under the curve is equal to 1 (100% of probability). We can now assume that the value of our continuous variable X depends on a very large number of other factors (in many cases beyond our possibility of direct analysis), the probability distribution of X becomes similar to a particular form of distribution, called normal distribution or Gauss distribution. The aforementioned concept is the famous Central Limit Theorem. The normal distribution represents a very important distribution of probability because f , that is the distribution of probability of our variables, can be represented by only two parameters:

An external file that holds a picture, illustration, etc.
Object name is jtd-07-03-E7-f1.jpg

Graphical description of the distribution of systolic blood pressure in a given population.

An external file that holds a picture, illustration, etc.
Object name is jtd-07-03-E7-f2.jpg

Graphical description of the normal distribution.

  • µ = mean;
  • σ = standard deviation.

The mean is a so-called measure of central tendency (it represents the more central value of our curve), while the standard deviation represents how dispersed are the values of probability around the central value (is a measure of dispersion).

  • The main characteristics of this distribution are:
  • It is symmetric around the µ ;
  • The area under the curve is 1;

If we consider the area under the curve between µ ± σ , this area will cover 68% of all the possible values of X , while the area between µ ± 2 σ , it will cover 95% of all the values.

The two parameters of the distribution are linked in the formula ( Eq. [10] ):

For µ = 0, and σ = 1, the curve is called standardized normal distribution. All the possible normal distributions of x may be “normalized” by defining a derived variable called z . ( Eq. [11] ):

To calculate the probability that our variable falls within a given interval, for instance z 0 and z 1 , we should calculate the following definite integral calculus ( Eq. [12] ):

Fortunately, for the standard normalized distribution of z every possible interval has been tabulated.

In a given population of adult men, the mean weight is 70 kg, with a standard deviation of 3 kg. What is the probability that a randomly selected individual from this population would have a weight of 65 kg or less?

To “normalize” our distribution, we should calculate the value of z ( Eq. [13] ):

Then, we should calculate the area under the curve ( Eq. [14] ):

The value of our interval has been already calculated and tabulated [the tables can be easily found in any text of statistics or in the web ( 4 )]. Our probability is 0.0475 (4.75%). We may also calculate the probability to find, within the same population, someone whose weight is between 65 and 74 kg. This probability can be seen as the difference of distribution between those whose weight is 74 kg or less and those whose weight is 65 kg or less: ( Eq. [15] ):

We already know that ( Eq. [16] ):

In the table we can find also the value for ( Eq. [17] ):

Our probability is ( Eq. [18] ):

Conclusions

The probability distributions are a common way to describe, and possibly predict, the probability of an event. The main point is to define the character of the variables whose behaviour we are trying to describe, trough probability (discrete or continuous). The identification of the right category will allow a proper application of a model (for instance, the standardized normal distribution) that would easily predict the probability of a given event.

Acknowledgements

Disclosure: The authors declare no conflict of interest.

research paper about statistics and probability

Academia.edu no longer supports Internet Explorer.

To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to  upgrade your browser .

  •  We're Hiring!
  •  Help Center

Probability and statistics

  • Most Cited Papers
  • Most Downloaded Papers
  • Newest Papers
  • Last »
  • Statistics Follow Following
  • Probability Theory Follow Following
  • Introduction to Probability Follow Following
  • Probability Follow Following
  • Applied Statistics Follow Following
  • Nonparametric Statistics Follow Following
  • Mathematics and Statistics Follow Following
  • Nonparametric Inference Follow Following
  • Nonparametric Econometrics Follow Following
  • Bayesian statistics & modelling Follow Following

Enter the email address you signed up with and we'll email you a reset link.

  • Academia.edu Journals
  •   We're Hiring!
  •   Help Center
  • Find new research papers in:
  • Health Sciences
  • Earth Sciences
  • Cognitive Science
  • Mathematics
  • Computer Science
  • Academia ©2024

Department of Statistics

Research topics in probability and statistics, problem solving in mathematics and statistics is inspiring and enjoyable. but are achievements in mathematics and statistics any of use in the so-called real world , researchers in the department of statistics at warwick are developing and utilising modern statistics, mathematics, and computing to solve practical problems., examples of themes for undergraduate research projects:.

  • Discovering which genes can discriminate between diseased and healthy patients
  • Modelling and detecting asset price bubbles while they are happening and before they burst
  • Modelling infectious diseases and identifying localized outbreaks
  • Developing a fast algorithm through probabilistic modeling for compression of sound data
  • Automatically diagnosing diseases with large-scale image data utilizing crime data for crime prevention and optimal allocation of police resources
  • Predicting the outcome of elections based on exit poll data
  • Computed Tomography validation of complex structures in Additive Layer Manufacturing

Probability of containment for multitype branching process models for emerging epidemics

Non-stationary statistical modeling and inference for circadian oscillations for research in cancer chronotherapy

Bayesian Models of Category-Specific Emotional Brain Responses

Decision focused inference on Networked Proabilistic Systems: with applications to food security

Rotationally invariant statistics for examining the evidence from the pores in fingerprints

Dynamic Uncertainty Handling for Coherent Decision Making in Nuclear Emergency Response

Study of Key Interventions into Terrorism using Bayesian Networks

Assessing the risk of subsequent tonic-clonic seizures in patients with a history of simple or complex partial seizures

Multidimensional Markov-functional Interest Rate Models

Prospect Theory, Liquidation and the Disposition Effect

Dynamic Bradley-Terry modelling of sports tournaments

Further information on the wide range of research opportunities open to you as an Undergraduate or Postgraduate Taught student in the Department of Statistics can be found on at our Student Research Opportunities webpage.

More information about research in the Department of Statistics, both applied and theoretical, can be found at the departmental research pages .

Mathematics as bridge

The work of mathematicians and statisticians often turns out useful and essential, but typically in a less concrete manner than say the work of a scientists or a physician. David Hilbert, in his now historical address to scientists and physicians, put it this way:

"The instrument that mediates between theory and practice, between thought and observation, is mathematics; it builds the connecting bridge and makes it stronger and stronger. Thus it happens that our entire present-day culture, insofar as it rests on intellectual insight into and harnessing of nature, is founded on mathematics"

Probability and Statistics in the 21st century

Almost a century after Hilbert's words, the mathematical fundations of sciences and social sciences, and the evidence based approach in medicine are often being taken for granted. In the 21st century we are facing complex big data sets with unknown structures from manifold aspecs of the 'real world' as well as fascinating discourses about objective and subjective notions of risk and uncertainty.

Probability and statistics are mathematical disciplines for modelling and analysing theoretical and practical aspects of these burning questions.

Have a language expert improve your writing

Run a free plagiarism check in 10 minutes, generate accurate citations for free.

  • Knowledge Base

The Beginner's Guide to Statistical Analysis | 5 Steps & Examples

Statistical analysis means investigating trends, patterns, and relationships using quantitative data . It is an important research tool used by scientists, governments, businesses, and other organizations.

To draw valid conclusions, statistical analysis requires careful planning from the very start of the research process . You need to specify your hypotheses and make decisions about your research design, sample size, and sampling procedure.

After collecting data from your sample, you can organize and summarize the data using descriptive statistics . Then, you can use inferential statistics to formally test hypotheses and make estimates about the population. Finally, you can interpret and generalize your findings.

This article is a practical introduction to statistical analysis for students and researchers. We’ll walk you through the steps using two research examples. The first investigates a potential cause-and-effect relationship, while the second investigates a potential correlation between variables.

Table of contents

Step 1: write your hypotheses and plan your research design, step 2: collect data from a sample, step 3: summarize your data with descriptive statistics, step 4: test hypotheses or make estimates with inferential statistics, step 5: interpret your results, other interesting articles.

To collect valid data for statistical analysis, you first need to specify your hypotheses and plan out your research design.

Writing statistical hypotheses

The goal of research is often to investigate a relationship between variables within a population . You start with a prediction, and use statistical analysis to test that prediction.

A statistical hypothesis is a formal way of writing a prediction about a population. Every research prediction is rephrased into null and alternative hypotheses that can be tested using sample data.

While the null hypothesis always predicts no effect or no relationship between variables, the alternative hypothesis states your research prediction of an effect or relationship.

  • Null hypothesis: A 5-minute meditation exercise will have no effect on math test scores in teenagers.
  • Alternative hypothesis: A 5-minute meditation exercise will improve math test scores in teenagers.
  • Null hypothesis: Parental income and GPA have no relationship with each other in college students.
  • Alternative hypothesis: Parental income and GPA are positively correlated in college students.

Planning your research design

A research design is your overall strategy for data collection and analysis. It determines the statistical tests you can use to test your hypothesis later on.

First, decide whether your research will use a descriptive, correlational, or experimental design. Experiments directly influence variables, whereas descriptive and correlational studies only measure variables.

  • In an experimental design , you can assess a cause-and-effect relationship (e.g., the effect of meditation on test scores) using statistical tests of comparison or regression.
  • In a correlational design , you can explore relationships between variables (e.g., parental income and GPA) without any assumption of causality using correlation coefficients and significance tests.
  • In a descriptive design , you can study the characteristics of a population or phenomenon (e.g., the prevalence of anxiety in U.S. college students) using statistical tests to draw inferences from sample data.

Your research design also concerns whether you’ll compare participants at the group level or individual level, or both.

  • In a between-subjects design , you compare the group-level outcomes of participants who have been exposed to different treatments (e.g., those who performed a meditation exercise vs those who didn’t).
  • In a within-subjects design , you compare repeated measures from participants who have participated in all treatments of a study (e.g., scores from before and after performing a meditation exercise).
  • In a mixed (factorial) design , one variable is altered between subjects and another is altered within subjects (e.g., pretest and posttest scores from participants who either did or didn’t do a meditation exercise).
  • Experimental
  • Correlational

First, you’ll take baseline test scores from participants. Then, your participants will undergo a 5-minute meditation exercise. Finally, you’ll record participants’ scores from a second math test.

In this experiment, the independent variable is the 5-minute meditation exercise, and the dependent variable is the math test score from before and after the intervention. Example: Correlational research design In a correlational study, you test whether there is a relationship between parental income and GPA in graduating college students. To collect your data, you will ask participants to fill in a survey and self-report their parents’ incomes and their own GPA.

Measuring variables

When planning a research design, you should operationalize your variables and decide exactly how you will measure them.

For statistical analysis, it’s important to consider the level of measurement of your variables, which tells you what kind of data they contain:

  • Categorical data represents groupings. These may be nominal (e.g., gender) or ordinal (e.g. level of language ability).
  • Quantitative data represents amounts. These may be on an interval scale (e.g. test score) or a ratio scale (e.g. age).

Many variables can be measured at different levels of precision. For example, age data can be quantitative (8 years old) or categorical (young). If a variable is coded numerically (e.g., level of agreement from 1–5), it doesn’t automatically mean that it’s quantitative instead of categorical.

Identifying the measurement level is important for choosing appropriate statistics and hypothesis tests. For example, you can calculate a mean score with quantitative data, but not with categorical data.

In a research study, along with measures of your variables of interest, you’ll often collect data on relevant participant characteristics.

Variable Type of data
Age Quantitative (ratio)
Gender Categorical (nominal)
Race or ethnicity Categorical (nominal)
Baseline test scores Quantitative (interval)
Final test scores Quantitative (interval)
Parental income Quantitative (ratio)
GPA Quantitative (interval)

Prevent plagiarism. Run a free check.

Population vs sample

In most cases, it’s too difficult or expensive to collect data from every member of the population you’re interested in studying. Instead, you’ll collect data from a sample.

Statistical analysis allows you to apply your findings beyond your own sample as long as you use appropriate sampling procedures . You should aim for a sample that is representative of the population.

Sampling for statistical analysis

There are two main approaches to selecting a sample.

  • Probability sampling: every member of the population has a chance of being selected for the study through random selection.
  • Non-probability sampling: some members of the population are more likely than others to be selected for the study because of criteria such as convenience or voluntary self-selection.

In theory, for highly generalizable findings, you should use a probability sampling method. Random selection reduces several types of research bias , like sampling bias , and ensures that data from your sample is actually typical of the population. Parametric tests can be used to make strong statistical inferences when data are collected using probability sampling.

But in practice, it’s rarely possible to gather the ideal sample. While non-probability samples are more likely to at risk for biases like self-selection bias , they are much easier to recruit and collect data from. Non-parametric tests are more appropriate for non-probability samples, but they result in weaker inferences about the population.

If you want to use parametric tests for non-probability samples, you have to make the case that:

  • your sample is representative of the population you’re generalizing your findings to.
  • your sample lacks systematic bias.

Keep in mind that external validity means that you can only generalize your conclusions to others who share the characteristics of your sample. For instance, results from Western, Educated, Industrialized, Rich and Democratic samples (e.g., college students in the US) aren’t automatically applicable to all non-WEIRD populations.

If you apply parametric tests to data from non-probability samples, be sure to elaborate on the limitations of how far your results can be generalized in your discussion section .

Create an appropriate sampling procedure

Based on the resources available for your research, decide on how you’ll recruit participants.

  • Will you have resources to advertise your study widely, including outside of your university setting?
  • Will you have the means to recruit a diverse sample that represents a broad population?
  • Do you have time to contact and follow up with members of hard-to-reach groups?

Your participants are self-selected by their schools. Although you’re using a non-probability sample, you aim for a diverse and representative sample. Example: Sampling (correlational study) Your main population of interest is male college students in the US. Using social media advertising, you recruit senior-year male college students from a smaller subpopulation: seven universities in the Boston area.

Calculate sufficient sample size

Before recruiting participants, decide on your sample size either by looking at other studies in your field or using statistics. A sample that’s too small may be unrepresentative of the sample, while a sample that’s too large will be more costly than necessary.

There are many sample size calculators online. Different formulas are used depending on whether you have subgroups or how rigorous your study should be (e.g., in clinical research). As a rule of thumb, a minimum of 30 units or more per subgroup is necessary.

To use these calculators, you have to understand and input these key components:

  • Significance level (alpha): the risk of rejecting a true null hypothesis that you are willing to take, usually set at 5%.
  • Statistical power : the probability of your study detecting an effect of a certain size if there is one, usually 80% or higher.
  • Expected effect size : a standardized indication of how large the expected result of your study will be, usually based on other similar studies.
  • Population standard deviation: an estimate of the population parameter based on a previous study or a pilot study of your own.

Once you’ve collected all of your data, you can inspect them and calculate descriptive statistics that summarize them.

Inspect your data

There are various ways to inspect your data, including the following:

  • Organizing data from each variable in frequency distribution tables .
  • Displaying data from a key variable in a bar chart to view the distribution of responses.
  • Visualizing the relationship between two variables using a scatter plot .

By visualizing your data in tables and graphs, you can assess whether your data follow a skewed or normal distribution and whether there are any outliers or missing data.

A normal distribution means that your data are symmetrically distributed around a center where most values lie, with the values tapering off at the tail ends.

Mean, median, mode, and standard deviation in a normal distribution

In contrast, a skewed distribution is asymmetric and has more values on one end than the other. The shape of the distribution is important to keep in mind because only some descriptive statistics should be used with skewed distributions.

Extreme outliers can also produce misleading statistics, so you may need a systematic approach to dealing with these values.

Calculate measures of central tendency

Measures of central tendency describe where most of the values in a data set lie. Three main measures of central tendency are often reported:

  • Mode : the most popular response or value in the data set.
  • Median : the value in the exact middle of the data set when ordered from low to high.
  • Mean : the sum of all values divided by the number of values.

However, depending on the shape of the distribution and level of measurement, only one or two of these measures may be appropriate. For example, many demographic characteristics can only be described using the mode or proportions, while a variable like reaction time may not have a mode at all.

Calculate measures of variability

Measures of variability tell you how spread out the values in a data set are. Four main measures of variability are often reported:

  • Range : the highest value minus the lowest value of the data set.
  • Interquartile range : the range of the middle half of the data set.
  • Standard deviation : the average distance between each value in your data set and the mean.
  • Variance : the square of the standard deviation.

Once again, the shape of the distribution and level of measurement should guide your choice of variability statistics. The interquartile range is the best measure for skewed distributions, while standard deviation and variance provide the best information for normal distributions.

Using your table, you should check whether the units of the descriptive statistics are comparable for pretest and posttest scores. For example, are the variance levels similar across the groups? Are there any extreme values? If there are, you may need to identify and remove extreme outliers in your data set or transform your data before performing a statistical test.

Pretest scores Posttest scores
Mean 68.44 75.25
Standard deviation 9.43 9.88
Variance 88.96 97.96
Range 36.25 45.12
30

From this table, we can see that the mean score increased after the meditation exercise, and the variances of the two scores are comparable. Next, we can perform a statistical test to find out if this improvement in test scores is statistically significant in the population. Example: Descriptive statistics (correlational study) After collecting data from 653 students, you tabulate descriptive statistics for annual parental income and GPA.

It’s important to check whether you have a broad range of data points. If you don’t, your data may be skewed towards some groups more than others (e.g., high academic achievers), and only limited inferences can be made about a relationship.

Parental income (USD) GPA
Mean 62,100 3.12
Standard deviation 15,000 0.45
Variance 225,000,000 0.16
Range 8,000–378,000 2.64–4.00
653

A number that describes a sample is called a statistic , while a number describing a population is called a parameter . Using inferential statistics , you can make conclusions about population parameters based on sample statistics.

Researchers often use two main methods (simultaneously) to make inferences in statistics.

  • Estimation: calculating population parameters based on sample statistics.
  • Hypothesis testing: a formal process for testing research predictions about the population using samples.

You can make two types of estimates of population parameters from sample statistics:

  • A point estimate : a value that represents your best guess of the exact parameter.
  • An interval estimate : a range of values that represent your best guess of where the parameter lies.

If your aim is to infer and report population characteristics from sample data, it’s best to use both point and interval estimates in your paper.

You can consider a sample statistic a point estimate for the population parameter when you have a representative sample (e.g., in a wide public opinion poll, the proportion of a sample that supports the current government is taken as the population proportion of government supporters).

There’s always error involved in estimation, so you should also provide a confidence interval as an interval estimate to show the variability around a point estimate.

A confidence interval uses the standard error and the z score from the standard normal distribution to convey where you’d generally expect to find the population parameter most of the time.

Hypothesis testing

Using data from a sample, you can test hypotheses about relationships between variables in the population. Hypothesis testing starts with the assumption that the null hypothesis is true in the population, and you use statistical tests to assess whether the null hypothesis can be rejected or not.

Statistical tests determine where your sample data would lie on an expected distribution of sample data if the null hypothesis were true. These tests give two main outputs:

  • A test statistic tells you how much your data differs from the null hypothesis of the test.
  • A p value tells you the likelihood of obtaining your results if the null hypothesis is actually true in the population.

Statistical tests come in three main varieties:

  • Comparison tests assess group differences in outcomes.
  • Regression tests assess cause-and-effect relationships between variables.
  • Correlation tests assess relationships between variables without assuming causation.

Your choice of statistical test depends on your research questions, research design, sampling method, and data characteristics.

Parametric tests

Parametric tests make powerful inferences about the population based on sample data. But to use them, some assumptions must be met, and only some types of variables can be used. If your data violate these assumptions, you can perform appropriate data transformations or use alternative non-parametric tests instead.

A regression models the extent to which changes in a predictor variable results in changes in outcome variable(s).

  • A simple linear regression includes one predictor variable and one outcome variable.
  • A multiple linear regression includes two or more predictor variables and one outcome variable.

Comparison tests usually compare the means of groups. These may be the means of different groups within a sample (e.g., a treatment and control group), the means of one sample group taken at different times (e.g., pretest and posttest scores), or a sample mean and a population mean.

  • A t test is for exactly 1 or 2 groups when the sample is small (30 or less).
  • A z test is for exactly 1 or 2 groups when the sample is large.
  • An ANOVA is for 3 or more groups.

The z and t tests have subtypes based on the number and types of samples and the hypotheses:

  • If you have only one sample that you want to compare to a population mean, use a one-sample test .
  • If you have paired measurements (within-subjects design), use a dependent (paired) samples test .
  • If you have completely separate measurements from two unmatched groups (between-subjects design), use an independent (unpaired) samples test .
  • If you expect a difference between groups in a specific direction, use a one-tailed test .
  • If you don’t have any expectations for the direction of a difference between groups, use a two-tailed test .

The only parametric correlation test is Pearson’s r . The correlation coefficient ( r ) tells you the strength of a linear relationship between two quantitative variables.

However, to test whether the correlation in the sample is strong enough to be important in the population, you also need to perform a significance test of the correlation coefficient, usually a t test, to obtain a p value. This test uses your sample size to calculate how much the correlation coefficient differs from zero in the population.

You use a dependent-samples, one-tailed t test to assess whether the meditation exercise significantly improved math test scores. The test gives you:

  • a t value (test statistic) of 3.00
  • a p value of 0.0028

Although Pearson’s r is a test statistic, it doesn’t tell you anything about how significant the correlation is in the population. You also need to test whether this sample correlation coefficient is large enough to demonstrate a correlation in the population.

A t test can also determine how significantly a correlation coefficient differs from zero based on sample size. Since you expect a positive correlation between parental income and GPA, you use a one-sample, one-tailed t test. The t test gives you:

  • a t value of 3.08
  • a p value of 0.001

Receive feedback on language, structure, and formatting

Professional editors proofread and edit your paper by focusing on:

  • Academic style
  • Vague sentences
  • Style consistency

See an example

research paper about statistics and probability

The final step of statistical analysis is interpreting your results.

Statistical significance

In hypothesis testing, statistical significance is the main criterion for forming conclusions. You compare your p value to a set significance level (usually 0.05) to decide whether your results are statistically significant or non-significant.

Statistically significant results are considered unlikely to have arisen solely due to chance. There is only a very low chance of such a result occurring if the null hypothesis is true in the population.

This means that you believe the meditation intervention, rather than random factors, directly caused the increase in test scores. Example: Interpret your results (correlational study) You compare your p value of 0.001 to your significance threshold of 0.05. With a p value under this threshold, you can reject the null hypothesis. This indicates a statistically significant correlation between parental income and GPA in male college students.

Note that correlation doesn’t always mean causation, because there are often many underlying factors contributing to a complex variable like GPA. Even if one variable is related to another, this may be because of a third variable influencing both of them, or indirect links between the two variables.

Effect size

A statistically significant result doesn’t necessarily mean that there are important real life applications or clinical outcomes for a finding.

In contrast, the effect size indicates the practical significance of your results. It’s important to report effect sizes along with your inferential statistics for a complete picture of your results. You should also report interval estimates of effect sizes if you’re writing an APA style paper .

With a Cohen’s d of 0.72, there’s medium to high practical significance to your finding that the meditation exercise improved test scores. Example: Effect size (correlational study) To determine the effect size of the correlation coefficient, you compare your Pearson’s r value to Cohen’s effect size criteria.

Decision errors

Type I and Type II errors are mistakes made in research conclusions. A Type I error means rejecting the null hypothesis when it’s actually true, while a Type II error means failing to reject the null hypothesis when it’s false.

You can aim to minimize the risk of these errors by selecting an optimal significance level and ensuring high power . However, there’s a trade-off between the two errors, so a fine balance is necessary.

Frequentist versus Bayesian statistics

Traditionally, frequentist statistics emphasizes null hypothesis significance testing and always starts with the assumption of a true null hypothesis.

However, Bayesian statistics has grown in popularity as an alternative approach in the last few decades. In this approach, you use previous research to continually update your hypotheses based on your expectations and observations.

Bayes factor compares the relative strength of evidence for the null versus the alternative hypothesis rather than making a conclusion about rejecting the null hypothesis or not.

If you want to know more about statistics , methodology , or research bias , make sure to check out some of our other articles with explanations and examples.

  • Student’s  t -distribution
  • Normal distribution
  • Null and Alternative Hypotheses
  • Chi square tests
  • Confidence interval

Methodology

  • Cluster sampling
  • Stratified sampling
  • Data cleansing
  • Reproducibility vs Replicability
  • Peer review
  • Likert scale

Research bias

  • Implicit bias
  • Framing effect
  • Cognitive bias
  • Placebo effect
  • Hawthorne effect
  • Hostile attribution bias
  • Affect heuristic

Is this article helpful?

Other students also liked.

  • Descriptive Statistics | Definitions, Types, Examples
  • Inferential Statistics | An Easy Introduction & Examples
  • Choosing the Right Statistical Test | Types & Examples

More interesting articles

  • Akaike Information Criterion | When & How to Use It (Example)
  • An Easy Introduction to Statistical Significance (With Examples)
  • An Introduction to t Tests | Definitions, Formula and Examples
  • ANOVA in R | A Complete Step-by-Step Guide with Examples
  • Central Limit Theorem | Formula, Definition & Examples
  • Central Tendency | Understanding the Mean, Median & Mode
  • Chi-Square (Χ²) Distributions | Definition & Examples
  • Chi-Square (Χ²) Table | Examples & Downloadable Table
  • Chi-Square (Χ²) Tests | Types, Formula & Examples
  • Chi-Square Goodness of Fit Test | Formula, Guide & Examples
  • Chi-Square Test of Independence | Formula, Guide & Examples
  • Coefficient of Determination (R²) | Calculation & Interpretation
  • Correlation Coefficient | Types, Formulas & Examples
  • Frequency Distribution | Tables, Types & Examples
  • How to Calculate Standard Deviation (Guide) | Calculator & Examples
  • How to Calculate Variance | Calculator, Analysis & Examples
  • How to Find Degrees of Freedom | Definition & Formula
  • How to Find Interquartile Range (IQR) | Calculator & Examples
  • How to Find Outliers | 4 Ways with Examples & Explanation
  • How to Find the Geometric Mean | Calculator & Formula
  • How to Find the Mean | Definition, Examples & Calculator
  • How to Find the Median | Definition, Examples & Calculator
  • How to Find the Mode | Definition, Examples & Calculator
  • How to Find the Range of a Data Set | Calculator & Formula
  • Hypothesis Testing | A Step-by-Step Guide with Easy Examples
  • Interval Data and How to Analyze It | Definitions & Examples
  • Levels of Measurement | Nominal, Ordinal, Interval and Ratio
  • Linear Regression in R | A Step-by-Step Guide & Examples
  • Missing Data | Types, Explanation, & Imputation
  • Multiple Linear Regression | A Quick Guide (Examples)
  • Nominal Data | Definition, Examples, Data Collection & Analysis
  • Normal Distribution | Examples, Formulas, & Uses
  • Null and Alternative Hypotheses | Definitions & Examples
  • One-way ANOVA | When and How to Use It (With Examples)
  • Ordinal Data | Definition, Examples, Data Collection & Analysis
  • Parameter vs Statistic | Definitions, Differences & Examples
  • Pearson Correlation Coefficient (r) | Guide & Examples
  • Poisson Distributions | Definition, Formula & Examples
  • Probability Distribution | Formula, Types, & Examples
  • Quartiles & Quantiles | Calculation, Definition & Interpretation
  • Ratio Scales | Definition, Examples, & Data Analysis
  • Simple Linear Regression | An Easy Introduction & Examples
  • Skewness | Definition, Examples & Formula
  • Statistical Power and Why It Matters | A Simple Introduction
  • Student's t Table (Free Download) | Guide & Examples
  • T-distribution: What it is and how to use it
  • Test statistics | Definition, Interpretation, and Examples
  • The Standard Normal Distribution | Calculator, Examples & Uses
  • Two-Way ANOVA | Examples & When To Use It
  • Type I & Type II Errors | Differences, Examples, Visualizations
  • Understanding Confidence Intervals | Easy Examples & Formulas
  • Understanding P values | Definition and Examples
  • Variability | Calculating Range, IQR, Variance, Standard Deviation
  • What is Effect Size and Why Does It Matter? (Examples)
  • What Is Kurtosis? | Definition, Examples & Formula
  • What Is Standard Error? | How to Calculate (Guide with Examples)

What is your plagiarism score?

StatAnalytica

Top 99+ Trending Statistics Research Topics for Students

statistics research topics

Being a statistics student, finding the best statistics research topics is quite challenging. But not anymore; find the best statistics research topics now!!!

Statistics is one of the tough subjects because it consists of lots of formulas, equations and many more. Therefore the students need to spend their time to understand these concepts. And when it comes to finding the best statistics research project for their topics, statistics students are always looking for someone to help them. 

In this blog, we will share with you the most interesting and trending statistics research topics in 2023. It will not just help you to stand out in your class but also help you to explore more about the world.

If you face any problem regarding statistics, then don’t worry. You can get the best statistics assignment help from one of our experts.

As you know, it is always suggested that you should work on interesting topics. That is why we have mentioned the most interesting research topics for college students and high school students. Here in this blog post, we will share with you the list of 99+ awesome statistics research topics.

Why Do We Need to Have Good Statistics Research Topics?

Table of Contents

Having a good research topic will not just help you score good grades, but it will also allow you to finish your project quickly. Because whenever we work on something interesting, our productivity automatically boosts. Thus, you need not invest lots of time and effort, and you can achieve the best with minimal effort and time. 

What Are Some Interesting Research Topics?

If we talk about the interesting research topics in statistics, it can vary from student to student. But here are the key topics that are quite interesting for almost every student:-

  • Literacy rate in a city.
  • Abortion and pregnancy rate in the USA.
  • Eating disorders in the citizens.
  • Parent role in self-esteem and confidence of the student.
  • Uses of AI in our daily life to business corporates.

Top 99+ Trending Statistics Research Topics For 2023

Here in this section, we will tell you more than 99 trending statistics research topics:

Sports Statistics Research Topics

  • Statistical analysis for legs and head injuries in Football.
  • Statistical analysis for shoulder and knee injuries in MotoGP.
  • Deep statistical evaluation for the doping test in sports from the past decade.
  • Statistical observation on the performance of athletes in the last Olympics.
  • Role and effect of sports in the life of the student.

Psychology Research Topics for Statistics

  • Deep statistical analysis of the effect of obesity on the student’s mental health in high school and college students.
  • Statistical evolution to find out the suicide reason among students and adults.
  • Statistics analysis to find out the effect of divorce on children in a country.
  • Psychology affects women because of the gender gap in specific country areas.
  • Statistics analysis to find out the cause of online bullying in students’ lives. 
  • In Psychology, PTSD and descriptive tendencies are discussed.
  • The function of researchers in statistical testing and probability.
  • Acceptable significance and probability thresholds in clinical Psychology.
  • The utilization of hypothesis and the role of P 0.05 for improved comprehension.
  • What types of statistical data are typically rejected in psychology?
  • The application of basic statistical principles and reasoning in psychological analysis.
  • The role of correlation is when several psychological concepts are at risk.
  • Actual case study learning and modeling are used to generate statistical reports.
  • In psychology, naturalistic observation is used as a research sample.
  • How should descriptive statistics be used to represent behavioral data sets?

Applied Statistics Research Topics

  • Does education have a deep impact on the financial success of an individual?
  • The investment in digital technology is having a meaningful return for corporations?
  • The gap of financial wealth between rich and poor in the USA.
  • A statistical approach to identify the effects of high-frequency trading in financial markets.
  • Statistics analysis to determine the impact of the multi-agent model in financial markets. 

Personalized Medicine Statistics Research Topics

  • Statistical analysis on the effect of methamphetamine on substance abusers.
  • Deep research on the impact of the Corona vaccine on the Omnicrone variant. 
  • Find out the best cancer treatment approach between orthodox therapies and alternative therapies.
  • Statistics analysis to identify the role of genes in the child’s overall immunity.
  • What factors help the patients to survive from Coronavirus .

Experimental Design Statistics Research Topics

  • Generic vs private education is one of the best for the students and has better financial return.
  • Psychology vs physiology: which leads the person not to quit their addictions?
  • Effect of breastmilk vs packed milk on the infant child overall development
  • Which causes more accidents: male alcoholics vs female alcoholics.
  • What causes the student not to reveal the cyberbullying in front of their parents in most cases. 

Easy Statistics Research Topics

  • Application of statistics in the world of data science
  • Statistics for finance: how statistics is helping the company to grow their finance
  • Advantages and disadvantages of Radar chart
  • Minor marriages in south-east Asia and African countries.
  • Discussion of ANOVA and correlation.
  • What statistical methods are most effective for active sports?
  • When measuring the correctness of college tests, a ranking statistical approach is used.
  • Statistics play an important role in Data Mining operations.
  • The practical application of heat estimation in engineering fields.
  • In the field of speech recognition, statistical analysis is used.
  • Estimating probiotics: how much time is necessary for an accurate statistical sample?
  • How will the United States population grow in the next twenty years?
  • The legislation and statistical reports deal with contentious issues.
  • The application of empirical entropy approaches with online grammar checking.
  • Transparency in statistical methodology and the reporting system of the United States Census Bureau.

Statistical Research Topics for High School

  • Uses of statistics in chemometrics
  • Statistics in business analytics and business intelligence
  • Importance of statistics in physics.
  • Deep discussion about multivariate statistics
  • Uses of Statistics in machine learning

Survey Topics for Statistics

  • Gather the data of the most qualified professionals in a specific area.
  • Survey the time wasted by the students in watching Tvs or Netflix.
  • Have a survey the fully vaccinated people in the USA 
  • Gather information on the effect of a government survey on the life of citizens
  • Survey to identify the English speakers in the world.

Statistics Research Paper Topics for Graduates

  • Have a deep decision of Bayes theorems
  • Discuss the Bayesian hierarchical models
  • Analysis of the process of Japanese restaurants. 
  • Deep analysis of Lévy’s continuity theorem
  • Analysis of the principle of maximum entropy

AP Statistics Topics

  • Discuss about the importance of econometrics
  • Analyze the pros and cons of Probit Model
  • Types of probability models and their uses
  • Deep discussion of ortho stochastic matrix
  • Find out the ways to get an adjacency matrix quickly

Good Statistics Research Topics 

  • National income and the regulation of cryptocurrency.
  • The benefits and drawbacks of regression analysis.
  • How can estimate methods be used to correct statistical differences?
  • Mathematical prediction models vs observation tactics.
  • In sociology research, there is bias in quantitative data analysis.
  • Inferential analytical approaches vs. descriptive statistics.
  • How reliable are AI-based methods in statistical analysis?
  • The internet news reporting and the fluctuations: statistics reports.
  • The importance of estimate in modeled statistics and artificial sampling.

Business Statistics Topics

  • Role of statistics in business in 2023
  • Importance of business statistics and analytics
  • What is the role of central tendency and dispersion in statistics
  • Best process of sampling business data.
  • Importance of statistics in big data.
  • The characteristics of business data sampling: benefits and cons of software solutions.
  • How may two different business tasks be tackled concurrently using linear regression analysis?
  • In economic data relations, index numbers, random probability, and correctness are all important.
  • The advantages of a dataset approach to statistics in programming statistics.
  • Commercial statistics: how should the data be prepared for maximum accuracy?

Statistical Research Topics for College Students

  • Evaluate the role of John Tukey’s contribution to statistics.
  • The role of statistics to improve ADHD treatment.
  • The uses and timeline of probability in statistics.
  • Deep analysis of Gertrude Cox’s experimental design in statistics.
  • Discuss about Florence Nightingale in statistics.
  • What sorts of music do college students prefer?
  • The Main Effect of Different Subjects on Student Performance.
  • The Importance of Analytics in Statistics Research.
  • The Influence of a Better Student in Class.
  • Do extracurricular activities help in the transformation of personalities?
  • Backbenchers’ Impact on Class Performance.
  • Medication’s Importance in Class Performance.
  • Are e-books better than traditional books?
  • Choosing aspects of a subject in college

How To Write Good Statistics Research Topics?

So, the main question that arises here is how you can write good statistics research topics. The trick is understanding the methodology that is used to collect and interpret statistical data. However, if you are trying to pick any topic for your statistics project, you must think about it before going any further. 

As a result, it will teach you about the data types that will be researched because the sample will be chosen correctly. On the other hand, your basic outline for choosing the correct topics is as follows:

  • Introduction of a problem
  • Methodology explanation and choice. 
  • Statistical research itself is in the main part (Body Part). 
  • Samples deviations and variables. 
  • Lastly, statistical interpretation is your last part (conclusion). 

Note:   Always include the sources from which you obtained the statistics data.

Top 3 Tips to Choose Good Statistics Research Topics

It can be quite easy for some students to pick a good statistics research topic without the help of an essay writer. But we know that it is not a common scenario for every student. That is why we will mention some of the best tips that will help you choose good statistics research topics for your next project. Either you are in a hurry or have enough time to explore. These tips will help you in every scenario.

1. Narrow down your research topic

We all start with many topics as we are not sure about our specific interests or niche. The initial step to picking up a good research topic for college or school students is to narrow down the research topic.

For this, you need to categorize the matter first. And then pick a specific category as per your interest. After that, brainstorm about the topic’s content and how you can make the points catchy, focused, directional, clear, and specific. 

2. Choose a topic that gives you curiosity

After categorizing the statistics research topics, it is time to pick one from the category. Don’t pick the most common topic because it will not help your grades and knowledge. Instead of it, please choose the best one, in which you have little information, or you are more likely to explore it.

In a statistics research paper, you always can explore something beyond your studies. By doing this, you will be more energetic to work on this project. And you will also feel glad to get them lots of information you were willing to have but didn’t get because of any reasons.

It will also make your professor happy to see your work. Ultimately it will affect your grades with a positive attitude.

3. Choose a manageable topic

Now you have decided on the topic, but you need to make sure that your research topic should be manageable. You will have limited time and resources to complete your project if you pick one of the deep statistics research topics with massive information.

Then you will struggle at the last moment and most probably not going to finish your project on time. Therefore, spend enough time exploring the topic and have a good idea about the time duration and resources you will use for the project. 

Statistics research topics are massive in numbers. Because statistics operations can be performed on anything from our psychology to our fitness. Therefore there are lots more statistics research topics to explore. But if you are not finding it challenging, then you can take the help of our statistics experts . They will help you to pick the most interesting and trending statistics research topics for your projects. 

With this help, you can also save your precious time to invest it in something else. You can also come up with a plethora of topics of your choice and we will help you to pick the best one among them. Apart from that, if you are working on a project and you are not sure whether that is the topic that excites you to work on it or not. Then we can also help you to clear all your doubts on the statistics research topic. 

Frequently Asked Questions

Q1. what are some good topics for the statistics project.

Have a look at some good topics for statistics projects:- 1. Research the average height and physics of basketball players. 2. Birth and death rate in a specific city or country. 3. Study on the obesity rate of children and adults in the USA. 4. The growth rate of China in the past few years 5. Major causes of injury in Football

Q2. What are the topics in statistics?

Statistics has lots of topics. It is hard to cover all of them in a short answer. But here are the major ones: conditional probability, variance, random variable, probability distributions, common discrete, and many more. 

Q3. What are the top 10 research topics?

Here are the top 10 research topics that you can try in 2023:

1. Plant Science 2. Mental health 3. Nutritional Immunology 4. Mood disorders 5. Aging brains 6. Infectious disease 7. Music therapy 8. Political misinformation 9. Canine Connection 10. Sustainable agriculture

Related Posts

how-to-find-the=best-online-statistics-homework-help

How to Find the Best Online Statistics Homework Help

why-spss-homework-help-is-an-important-aspects-for-students

Why SPSS Homework Help Is An Important aspect for Students?

Stochastic Stability Analysis of Ship Rolling Based on the Stochastic Averaging Method and First Passage Probability

28 Pages Posted: 19 Sep 2024

Tianjin University

Jianxing Yu

Osaka University

China Ship Scientific Research Center

Hamburg University of Technology

Prediction of dynamics stability of ships under complex sea state is a major scientific issue in the field of safe navigation. Stability discrimination methods based on deterministic waves cannot accurately identify the instability domain in random waves. In this paper, the stability characteristics of nonlinear roll motion in random waves are investigated based on stochastic analysis method. The stability boundary of roll system with only parametric excitation is given based on the stochastic averaging method. The stochastic Hopf bifurcation phenomenon of is demonstrated. Based on the first passage theory, the first passage probability of the stochastic roll motion is calculated with given boundary conditions and initial conditions. The entire sea areas are divided into three parts, including high stability, medium stability, and low stability regions. Based on the first passage probability approach, the probabilities of the roll response exceeding 25° are calculated for different wave directions. The results of the two methods are compared and the pros and cons of using the two methods to assess the stability of large roll are analyzed. With the two methods combined, more information for stochastic stability of roll motion under different wave directions can be provided.

Keywords: Nonlinear roll, Additive excitation, Stochastic stability, Stochastic averaging method, First passage probability

Suggested Citation: Suggested Citation

Tianjin University ( email )

92, Weijin Road Nankai District Tianjin, 300072 China

Liqin Liu (Contact Author)

Osaka university ( email ).

1-1 Yamadaoka Suita Osaka, 565-0871 Japan

China Ship Scientific Research Center ( email )

Hamburg university of technology ( email ).

Hamburg Germany

Do you have a job opening that you would like to promote on SSRN?

Paper statistics, related ejournals, stochastic models ejournal.

Subscribe to this fee journal for more curated articles on this topic

Aerospace Engineering eJournal

Difference & differential equations ejournal, dynamical systems ejournal, computational physics ejournal.

COMMENTS

  1. Journal of Probability and Statistics

    Print ISSN: 1687-952X. Journal of Probability and Statistics is an open access journal publishing papers on the theory and application of probability and statistics that consider new methods and approaches to their implementation, or report significant results for the field. As part of Wiley's Forward Series, this journal offers a streamlined ...

  2. Statistics & Probability Letters

    About the journal. Statistics & Probability Letters adopts a novel and highly innovative approach to the publication of research findings in statistics and probability. It features concise articles, rapid publication and broad coverage of the statistics and probability literature. Statistics & Probability Letters is a refereed journal.

  3. Introduction to Research Statistical Analysis: An Overview of the

    Introduction. Statistical analysis is necessary for any research project seeking to make quantitative conclusions. The following is a primer for research-based statistical analysis. It is intended to be a high-level overview of appropriate statistical testing, while not diving too deep into any specific methodology.

  4. Probability and Statistics

    Feature papers represent the most advanced research with significant potential for high impact in the field. A Feature Paper should be a substantial original Article that involves several techniques or approaches, provides an outlook for future research directions and describes possible research applications. ... The section Probability and ...

  5. Statistics

    A central tenet of statistics is to describe the variations in a data set or population using probability distributions. ... effects of biases in research and services that are dependent on ...

  6. Probability, Statistics and Their Applications 2021

    This paper focuses mainly on the problem of computing the γth, γ> 0, moment of a random variable Yn: = ∑n i = 1αiXi in which the αi 's are positive [...] Read more. (This article belongs to the Special Issue Probability, Statistics and Their Applications 2021) Show Figures. 16 pages, 8765 KiB.

  7. (PDF) Teaching and Learning of Probability

    Abstract. In this paper, we dev elop a personal synthesis of the most outstanding research on the teaching and learning of probability. in the past years. W e conducted a systematic search to ex ...

  8. Research in Statistics

    Taylor & Francis are currently supporting a 100% APC discount for all authors. Research in Statistics is a broad open access journal publishing original research in all areas of statistics and probability.The journal focuses on broadening existing research fields, and in facilitating international collaboration, and is devoted to the international advancement of the theory and application of ...

  9. 112130 PDFs

    The analysis of random phenomena and variable, stochastic processes and modeling. | Explore the latest full-text research PDFs, articles, conference papers, preprints and more on PROBABILITY THEORY.

  10. Mathematics

    Feature papers represent the most advanced research with significant potential for high impact in the field. A Feature Paper should be a substantial original Article that involves several techniques or approaches, provides an outlook for future research directions and describes possible research applications. ... Statistics and probability are ...

  11. Statistics and Probability Have Always Been Value-Laden: An Historical

    Quantitative researchers often discuss research ethics as if specific ethical problems can be reduced to abstract normative logics (e.g., virtue ethics, utilitarianism, deontology). Such approaches overlook how values are embedded in every aspect of quantitative methods, including 'observations,' 'facts,' and notions of 'objectivity.' We describe how quantitative research practices ...

  12. Difficulties in Learning Basic Concepts in Probability and Statistics

    In this paper we review the literature related to difficulties that students. at the precollege and college level have in understanding probability and statistics and discuss the research that is beginning to illuminate the exact nature of the difficulties, their causes, and what can be done about them.

  13. Undergraduate Research in an Applied Probability & Statistics Course

    ABSTRACT: The author modi ed a course entitled Applied Probability and Statistics to engage students in performing their own hypothesis tests in a course-based undergraduate research experience (CURE). This paper discusses the structure of the course, what it does to encourage undergraduate re-search, and changes one could make to tailor this ...

  14. A practical overview on probability distributions

    A short definition of probability. We can define the probability of a given event by evaluating, in previous observations, the incidence of the same event under circumstances that are as similar as possible to the circumstances we are observing [this is the frequentistic definition of probability, and is based on the relative frequency of an observed event, observed in previous circumstances ()].

  15. Introduction to the Concept of Likelihood and Its Applications

    Likelihood is a strange concept in that it is not a probability but is proportional to a probability. The likelihood of a hypothesis (H) given some data (D) is the probability of obtaining D given that H is true multiplied by an arbitrary positive constant K: L(H) = K × P(D|H).In most cases, a hypothesis represents a value of a parameter in a statistical model, such as the mean of a normal ...

  16. PDF Student S Attitudes Towards Probability and Statistics and Academic

    he productivity or benefits that Probability and Statistics can offer. Items: 1, 11, 15, 21 and 24.The items of this factor that most reflect this aspect are the following: (1) 85.8% of students agree that Statistics and Probability are much needed subjects in their careers and (2) 63.5% of students ag. ee.

  17. (PDF) Probability and Statistics

    The chapter is written for undergraduate and graduate students interested in probability and statistics, as well as for mathematicians working in other areas of mathematics, who would like to ...

  18. Statistics and Probability Research Papers

    Data Analysis, Complex network, Empirical Evaluation, Statistics and Probability Characterizing and modeling cyclic behavior in non-stationary time series through multi-resolution analysis A method based on wavelet transform and genetic programming is proposed for characterizing and modeling variations at multiple scales in non-stationary time ...

  19. Probability and statistics Research Papers

    Data Science as a scientific discipline is influenced by informatics, computer science, mathematics, operations research, and statistics as well as the applied sciences. Many research papers have come across that probability and statistics are the crucial prerequisites for data science.

  20. Research topics in Probability and Statistics

    Examples of themes for undergraduate research projects: Discovering which genes can discriminate between diseased and healthy patients. Modelling and detecting asset price bubbles while they are happening and before they burst. Modelling infectious diseases and identifying localized outbreaks. Developing a fast algorithm through probabilistic ...

  21. The Beginner's Guide to Statistical Analysis

    Table of contents. Step 1: Write your hypotheses and plan your research design. Step 2: Collect data from a sample. Step 3: Summarize your data with descriptive statistics. Step 4: Test hypotheses or make estimates with inferential statistics.

  22. Key Statistical Publications You Should Know

    Wald's paper focuses on making decisions in statistics. He introduces the idea of a decision function, which guides how to make decisions based on data. Wald also discusses making optimal choices by weighing the importance of different outcomes. 3. An Introduction to the Bootstrap, by B. Efron and R.J. Tibshirani (1993)

  23. Top 99+ Trending Statistics Research Topics for Students

    If we talk about the interesting research topics in statistics, it can vary from student to student. But here are the key topics that are quite interesting for almost every student:-. Literacy rate in a city. Abortion and pregnancy rate in the USA. Eating disorders in the citizens.

  24. Statistics and Probability

    Unit 1. Individuals, variables, and categorical & quantitative data. Unit 2. Displaying and comparing quantitative data: Unit test. Unit 3. Effects of shifting, adding, & removing a data point. Unit 4. Unit 5. Positive and negative linear associations from scatter plots.

  25. 2.5: Scientific method and where statistics fits

    When we say "experiment," the hypothesis came first. More commonly in statistics, the phrases planned, and therefore a priori, and unplanned or a posterori, comparisons are referenced.In practice, biologists design experiments and make observations accordingly to test one or more hypotheses, but they may also address additional hypotheses after the fact, especially if the experiment ...

  26. Stochastic Stability Analysis of Ship Rolling Based on the Stochastic

    In this paper, the stability characteristics of nonlinear roll motion in random waves are investigated based on stochastic analysis method. ... Based on the first passage probability approach, the probabilities of the roll response exceeding 25° are calculated for different wave directions. The results of the two methods are compared and the ...