example of statistical analysis research paper

Statistical Papers

Statistical Papers is a forum for presentation and critical assessment of statistical methods encouraging the discussion of methodological foundations and potential applications.

  • The Journal stresses statistical methods that have broad applications, giving special attention to those relevant to the economic and social sciences.
  • Covers all topics of modern data science, such as frequentist and Bayesian design and inference as well as statistical learning.
  • Contains original research papers (regular articles), survey articles, short communications, reports on statistical software, and book reviews.
  • High author satisfaction with 90% likely to publish in the journal again.
  • Werner G. Müller,
  • Carsten Jentsch,
  • Shuangzhe Liu,
  • Ulrike Schneider

example of statistical analysis research paper

Latest issue

Volume 65, Issue 2

Latest articles

On some stable linear functional regression estimators based on random projections.

  • Asma Ben Saber
  • Abderrazek Karoui

example of statistical analysis research paper

Testing practical relevance of treatment effects

  • Andrea Ongaro
  • Sonia Migliorati
  • Enrico Ripamonti

example of statistical analysis research paper

Supervised dimension reduction for functional time series

  • Guochang Wang
  • Zengyao Wen
  • Shanshan Liang

example of statistical analysis research paper

Statistical inferences for missing response problems based on modified empirical likelihood

  • Sima Sharghi
  • Kevin Stoll

example of statistical analysis research paper

A high-dimensional single-index regression for interactions between treatment and covariates

  • Thaddeus Tarpey
  • R. Todd Ogden

example of statistical analysis research paper

Journal updates

Write & submit: overleaf latex template.

Overleaf LaTeX Template

Journal information

  • Australian Business Deans Council (ABDC) Journal Quality List
  • Current Index to Statistics
  • Google Scholar
  • Japanese Science and Technology Agency (JST)
  • Mathematical Reviews
  • Norwegian Register for Scientific Journals and Series
  • OCLC WorldCat Discovery Service
  • Research Papers in Economics (RePEc)
  • Science Citation Index Expanded (SCIE)
  • TD Net Discovery Service
  • UGC-CARE List (India)

Rights and permissions

Springer policies

© Springer-Verlag GmbH Germany, part of Springer Nature

  • Find a journal
  • Publish with us
  • Track your research

Table of Contents

Types of statistical analysis, importance of statistical analysis, benefits of statistical analysis, statistical analysis process, statistical analysis methods, statistical analysis software, statistical analysis examples, career in statistical analysis, choose the right program, become proficient in statistics today, what is statistical analysis types, methods and examples.

What Is Statistical Analysis?

Statistical analysis is the process of collecting and analyzing data in order to discern patterns and trends. It is a method for removing bias from evaluating data by employing numerical analysis. This technique is useful for collecting the interpretations of research, developing statistical models, and planning surveys and studies.

Statistical analysis is a scientific tool in AI and ML that helps collect and analyze large amounts of data to identify common patterns and trends to convert them into meaningful information. In simple words, statistical analysis is a data analysis tool that helps draw meaningful conclusions from raw and unstructured data. 

The conclusions are drawn using statistical analysis facilitating decision-making and helping businesses make future predictions on the basis of past trends. It can be defined as a science of collecting and analyzing data to identify trends and patterns and presenting them. Statistical analysis involves working with numbers and is used by businesses and other institutions to make use of data to derive meaningful information. 

Given below are the 6 types of statistical analysis:

Descriptive Analysis

Descriptive statistical analysis involves collecting, interpreting, analyzing, and summarizing data to present them in the form of charts, graphs, and tables. Rather than drawing conclusions, it simply makes the complex data easy to read and understand.

Inferential Analysis

The inferential statistical analysis focuses on drawing meaningful conclusions on the basis of the data analyzed. It studies the relationship between different variables or makes predictions for the whole population.

Predictive Analysis

Predictive statistical analysis is a type of statistical analysis that analyzes data to derive past trends and predict future events on the basis of them. It uses machine learning algorithms, data mining , data modelling , and artificial intelligence to conduct the statistical analysis of data.

Prescriptive Analysis

The prescriptive analysis conducts the analysis of data and prescribes the best course of action based on the results. It is a type of statistical analysis that helps you make an informed decision. 

Exploratory Data Analysis

Exploratory analysis is similar to inferential analysis, but the difference is that it involves exploring the unknown data associations. It analyzes the potential relationships within the data. 

Causal Analysis

The causal statistical analysis focuses on determining the cause and effect relationship between different variables within the raw data. In simple words, it determines why something happens and its effect on other variables. This methodology can be used by businesses to determine the reason for failure. 

Statistical analysis eliminates unnecessary information and catalogs important data in an uncomplicated manner, making the monumental work of organizing inputs appear so serene. Once the data has been collected, statistical analysis may be utilized for a variety of purposes. Some of them are listed below:

  • The statistical analysis aids in summarizing enormous amounts of data into clearly digestible chunks.
  • The statistical analysis aids in the effective design of laboratory, field, and survey investigations.
  • Statistical analysis may help with solid and efficient planning in any subject of study.
  • Statistical analysis aid in establishing broad generalizations and forecasting how much of something will occur under particular conditions.
  • Statistical methods, which are effective tools for interpreting numerical data, are applied in practically every field of study. Statistical approaches have been created and are increasingly applied in physical and biological sciences, such as genetics.
  • Statistical approaches are used in the job of a businessman, a manufacturer, and a researcher. Statistics departments can be found in banks, insurance businesses, and government agencies.
  • A modern administrator, whether in the public or commercial sector, relies on statistical data to make correct decisions.
  • Politicians can utilize statistics to support and validate their claims while also explaining the issues they address.

Become a Data Science & Business Analytics Professional

  • 28% Annual Job Growth By 2026
  • 11.5 M Expected New Jobs For Data Science By 2026

Data Analyst

  • Industry-recognized Data Analyst Master’s certificate from Simplilearn
  • Dedicated live sessions by faculty of industry experts

Data Scientist

  • Add the IBM Advantage to your Learning
  • 25 Industry-relevant Projects and Integrated labs

Here's what learners are saying regarding our programs:

Gayathri Ramesh

Gayathri Ramesh

Associate data engineer , publicis sapient.

The course was well structured and curated. The live classes were extremely helpful. They made learning more productive and interactive. The program helped me change my domain from a data analyst to an Associate Data Engineer.

A.Anthony Davis

A.Anthony Davis

Simplilearn has one of the best programs available online to earn real-world skills that are in demand worldwide. I just completed the Machine Learning Advanced course, and the LMS was excellent.

Statistical analysis can be called a boon to mankind and has many benefits for both individuals and organizations. Given below are some of the reasons why you should consider investing in statistical analysis:

  • It can help you determine the monthly, quarterly, yearly figures of sales profits, and costs making it easier to make your decisions.
  • It can help you make informed and correct decisions.
  • It can help you identify the problem or cause of the failure and make corrections. For example, it can identify the reason for an increase in total costs and help you cut the wasteful expenses.
  • It can help you conduct market analysis and make an effective marketing and sales strategy.
  • It helps improve the efficiency of different processes.

Given below are the 5 steps to conduct a statistical analysis that you should follow:

  • Step 1: Identify and describe the nature of the data that you are supposed to analyze.
  • Step 2: The next step is to establish a relation between the data analyzed and the sample population to which the data belongs. 
  • Step 3: The third step is to create a model that clearly presents and summarizes the relationship between the population and the data.
  • Step 4: Prove if the model is valid or not.
  • Step 5: Use predictive analysis to predict future trends and events likely to happen. 

Although there are various methods used to perform data analysis, given below are the 5 most used and popular methods of statistical analysis:

Mean or average mean is one of the most popular methods of statistical analysis. Mean determines the overall trend of the data and is very simple to calculate. Mean is calculated by summing the numbers in the data set together and then dividing it by the number of data points. Despite the ease of calculation and its benefits, it is not advisable to resort to mean as the only statistical indicator as it can result in inaccurate decision making. 

Standard Deviation

Standard deviation is another very widely used statistical tool or method. It analyzes the deviation of different data points from the mean of the entire data set. It determines how data of the data set is spread around the mean. You can use it to decide whether the research outcomes can be generalized or not. 

Regression is a statistical tool that helps determine the cause and effect relationship between the variables. It determines the relationship between a dependent and an independent variable. It is generally used to predict future trends and events.

Hypothesis Testing

Hypothesis testing can be used to test the validity or trueness of a conclusion or argument against a data set. The hypothesis is an assumption made at the beginning of the research and can hold or be false based on the analysis results. 

Sample Size Determination

Sample size determination or data sampling is a technique used to derive a sample from the entire population, which is representative of the population. This method is used when the size of the population is very large. You can choose from among the various data sampling techniques such as snowball sampling, convenience sampling, and random sampling. 

Everyone can't perform very complex statistical calculations with accuracy making statistical analysis a time-consuming and costly process. Statistical software has become a very important tool for companies to perform their data analysis. The software uses Artificial Intelligence and Machine Learning to perform complex calculations, identify trends and patterns, and create charts, graphs, and tables accurately within minutes. 

Look at the standard deviation sample calculation given below to understand more about statistical analysis.

The weights of 5 pizza bases in cms are as follows:

Calculation of Mean = (9+2+5+4+12)/5 = 32/5 = 6.4

Calculation of mean of squared mean deviation = (6.76+19.36+1.96+5.76+31.36)/5 = 13.04

Sample Variance = 13.04

Standard deviation = √13.04 = 3.611

A Statistical Analyst's career path is determined by the industry in which they work. Anyone interested in becoming a Data Analyst may usually enter the profession and qualify for entry-level Data Analyst positions right out of high school or a certificate program — potentially with a Bachelor's degree in statistics, computer science, or mathematics. Some people go into data analysis from a similar sector such as business, economics, or even the social sciences, usually by updating their skills mid-career with a statistical analytics course.

Statistical Analyst is also a great way to get started in the normally more complex area of data science. A Data Scientist is generally a more senior role than a Data Analyst since it is more strategic in nature and necessitates a more highly developed set of technical abilities, such as knowledge of multiple statistical tools, programming languages, and predictive analytics models.

Aspiring Data Scientists and Statistical Analysts generally begin their careers by learning a programming language such as R or SQL. Following that, they must learn how to create databases, do basic analysis, and make visuals using applications such as Tableau. However, not every Statistical Analyst will need to know how to do all of these things, but if you want to advance in your profession, you should be able to do them all.

Based on your industry and the sort of work you do, you may opt to study Python or R, become an expert at data cleaning, or focus on developing complicated statistical models.

You could also learn a little bit of everything, which might help you take on a leadership role and advance to the position of Senior Data Analyst. A Senior Statistical Analyst with vast and deep knowledge might take on a leadership role leading a team of other Statistical Analysts. Statistical Analysts with extra skill training may be able to advance to Data Scientists or other more senior data analytics positions.

Supercharge your career in AI and ML with Simplilearn's comprehensive courses. Gain the skills and knowledge to transform industries and unleash your true potential. Enroll now and unlock limitless possibilities!

Program Name AI Engineer Post Graduate Program In Artificial Intelligence Post Graduate Program In Artificial Intelligence Geo All Geos All Geos IN/ROW University Simplilearn Purdue Caltech Course Duration 11 Months 11 Months 11 Months Coding Experience Required Basic Basic No Skills You Will Learn 10+ skills including data structure, data manipulation, NumPy, Scikit-Learn, Tableau and more. 16+ skills including chatbots, NLP, Python, Keras and more. 8+ skills including Supervised & Unsupervised Learning Deep Learning Data Visualization, and more. Additional Benefits Get access to exclusive Hackathons, Masterclasses and Ask-Me-Anything sessions by IBM Applied learning via 3 Capstone and 12 Industry-relevant Projects Purdue Alumni Association Membership Free IIMJobs Pro-Membership of 6 months Resume Building Assistance Upto 14 CEU Credits Caltech CTME Circle Membership Cost $$ $$$$ $$$$ Explore Program Explore Program Explore Program

Hope this article assisted you in understanding the importance of statistical analysis in every sphere of life. Artificial Intelligence (AI) can help you perform statistical analysis and data analysis very effectively and efficiently. 

If you are a science wizard and fascinated by the role of AI in statistical analysis, check out this amazing Caltech Post Graduate Program in AI & ML course in collaboration with Caltech. With a comprehensive syllabus and real-life projects, this course is one of the most popular courses and will help you with all that you need to know about Artificial Intelligence. 

Our AI & Machine Learning Courses Duration And Fees

AI & Machine Learning Courses typically range from a few weeks to several months, with fees varying based on program and institution.

Get Free Certifications with free video courses

Introduction to Data Analytics Course

Data Science & Business Analytics

Introduction to Data Analytics Course

Introduction to Data Science

Introduction to Data Science

Learn from Industry Experts with free Masterclasses

Ai & machine learning.

Career Masterclass: How to Build the Best Fantasy League Team Using Gen AI Tools

Gain Gen AI expertise in Purdue's Applied Gen AI Specialization

Unlock Your Career Potential: Land Your Dream Job with Gen AI Tools

Recommended Reads

Free eBook: Guide To The CCBA And CBAP Certifications

Understanding Statistical Process Control (SPC) and Top Applications

A Complete Guide on the Types of Statistical Studies

Digital Marketing Salary Guide 2021

What Is Data Analysis: A Comprehensive Guide

A Complete Guide to Get a Grasp of Time Series Analysis

Get Affiliated Certifications with Live Class programs

  • PMP, PMI, PMBOK, CAPM, PgMP, PfMP, ACP, PBA, RMP, SP, and OPM3 are registered marks of the Project Management Institute, Inc.

Have a thesis expert improve your writing

Check your thesis for plagiarism in 10 minutes, generate your apa citations for free.

  • Knowledge Base

The Beginner's Guide to Statistical Analysis | 5 Steps & Examples

Statistical analysis means investigating trends, patterns, and relationships using quantitative data . It is an important research tool used by scientists, governments, businesses, and other organisations.

To draw valid conclusions, statistical analysis requires careful planning from the very start of the research process . You need to specify your hypotheses and make decisions about your research design, sample size, and sampling procedure.

After collecting data from your sample, you can organise and summarise the data using descriptive statistics . Then, you can use inferential statistics to formally test hypotheses and make estimates about the population. Finally, you can interpret and generalise your findings.

This article is a practical introduction to statistical analysis for students and researchers. We’ll walk you through the steps using two research examples. The first investigates a potential cause-and-effect relationship, while the second investigates a potential correlation between variables.

Table of contents

Step 1: write your hypotheses and plan your research design, step 2: collect data from a sample, step 3: summarise your data with descriptive statistics, step 4: test hypotheses or make estimates with inferential statistics, step 5: interpret your results, frequently asked questions about statistics.

To collect valid data for statistical analysis, you first need to specify your hypotheses and plan out your research design.

Writing statistical hypotheses

The goal of research is often to investigate a relationship between variables within a population . You start with a prediction, and use statistical analysis to test that prediction.

A statistical hypothesis is a formal way of writing a prediction about a population. Every research prediction is rephrased into null and alternative hypotheses that can be tested using sample data.

While the null hypothesis always predicts no effect or no relationship between variables, the alternative hypothesis states your research prediction of an effect or relationship.

  • Null hypothesis: A 5-minute meditation exercise will have no effect on math test scores in teenagers.
  • Alternative hypothesis: A 5-minute meditation exercise will improve math test scores in teenagers.
  • Null hypothesis: Parental income and GPA have no relationship with each other in college students.
  • Alternative hypothesis: Parental income and GPA are positively correlated in college students.

Planning your research design

A research design is your overall strategy for data collection and analysis. It determines the statistical tests you can use to test your hypothesis later on.

First, decide whether your research will use a descriptive, correlational, or experimental design. Experiments directly influence variables, whereas descriptive and correlational studies only measure variables.

  • In an experimental design , you can assess a cause-and-effect relationship (e.g., the effect of meditation on test scores) using statistical tests of comparison or regression.
  • In a correlational design , you can explore relationships between variables (e.g., parental income and GPA) without any assumption of causality using correlation coefficients and significance tests.
  • In a descriptive design , you can study the characteristics of a population or phenomenon (e.g., the prevalence of anxiety in U.S. college students) using statistical tests to draw inferences from sample data.

Your research design also concerns whether you’ll compare participants at the group level or individual level, or both.

  • In a between-subjects design , you compare the group-level outcomes of participants who have been exposed to different treatments (e.g., those who performed a meditation exercise vs those who didn’t).
  • In a within-subjects design , you compare repeated measures from participants who have participated in all treatments of a study (e.g., scores from before and after performing a meditation exercise).
  • In a mixed (factorial) design , one variable is altered between subjects and another is altered within subjects (e.g., pretest and posttest scores from participants who either did or didn’t do a meditation exercise).
  • Experimental
  • Correlational

First, you’ll take baseline test scores from participants. Then, your participants will undergo a 5-minute meditation exercise. Finally, you’ll record participants’ scores from a second math test.

In this experiment, the independent variable is the 5-minute meditation exercise, and the dependent variable is the math test score from before and after the intervention. Example: Correlational research design In a correlational study, you test whether there is a relationship between parental income and GPA in graduating college students. To collect your data, you will ask participants to fill in a survey and self-report their parents’ incomes and their own GPA.

Measuring variables

When planning a research design, you should operationalise your variables and decide exactly how you will measure them.

For statistical analysis, it’s important to consider the level of measurement of your variables, which tells you what kind of data they contain:

  • Categorical data represents groupings. These may be nominal (e.g., gender) or ordinal (e.g. level of language ability).
  • Quantitative data represents amounts. These may be on an interval scale (e.g. test score) or a ratio scale (e.g. age).

Many variables can be measured at different levels of precision. For example, age data can be quantitative (8 years old) or categorical (young). If a variable is coded numerically (e.g., level of agreement from 1–5), it doesn’t automatically mean that it’s quantitative instead of categorical.

Identifying the measurement level is important for choosing appropriate statistics and hypothesis tests. For example, you can calculate a mean score with quantitative data, but not with categorical data.

In a research study, along with measures of your variables of interest, you’ll often collect data on relevant participant characteristics.

Population vs sample

In most cases, it’s too difficult or expensive to collect data from every member of the population you’re interested in studying. Instead, you’ll collect data from a sample.

Statistical analysis allows you to apply your findings beyond your own sample as long as you use appropriate sampling procedures . You should aim for a sample that is representative of the population.

Sampling for statistical analysis

There are two main approaches to selecting a sample.

  • Probability sampling: every member of the population has a chance of being selected for the study through random selection.
  • Non-probability sampling: some members of the population are more likely than others to be selected for the study because of criteria such as convenience or voluntary self-selection.

In theory, for highly generalisable findings, you should use a probability sampling method. Random selection reduces sampling bias and ensures that data from your sample is actually typical of the population. Parametric tests can be used to make strong statistical inferences when data are collected using probability sampling.

But in practice, it’s rarely possible to gather the ideal sample. While non-probability samples are more likely to be biased, they are much easier to recruit and collect data from. Non-parametric tests are more appropriate for non-probability samples, but they result in weaker inferences about the population.

If you want to use parametric tests for non-probability samples, you have to make the case that:

  • your sample is representative of the population you’re generalising your findings to.
  • your sample lacks systematic bias.

Keep in mind that external validity means that you can only generalise your conclusions to others who share the characteristics of your sample. For instance, results from Western, Educated, Industrialised, Rich and Democratic samples (e.g., college students in the US) aren’t automatically applicable to all non-WEIRD populations.

If you apply parametric tests to data from non-probability samples, be sure to elaborate on the limitations of how far your results can be generalised in your discussion section .

Create an appropriate sampling procedure

Based on the resources available for your research, decide on how you’ll recruit participants.

  • Will you have resources to advertise your study widely, including outside of your university setting?
  • Will you have the means to recruit a diverse sample that represents a broad population?
  • Do you have time to contact and follow up with members of hard-to-reach groups?

Your participants are self-selected by their schools. Although you’re using a non-probability sample, you aim for a diverse and representative sample. Example: Sampling (correlational study) Your main population of interest is male college students in the US. Using social media advertising, you recruit senior-year male college students from a smaller subpopulation: seven universities in the Boston area.

Calculate sufficient sample size

Before recruiting participants, decide on your sample size either by looking at other studies in your field or using statistics. A sample that’s too small may be unrepresentative of the sample, while a sample that’s too large will be more costly than necessary.

There are many sample size calculators online. Different formulas are used depending on whether you have subgroups or how rigorous your study should be (e.g., in clinical research). As a rule of thumb, a minimum of 30 units or more per subgroup is necessary.

To use these calculators, you have to understand and input these key components:

  • Significance level (alpha): the risk of rejecting a true null hypothesis that you are willing to take, usually set at 5%.
  • Statistical power : the probability of your study detecting an effect of a certain size if there is one, usually 80% or higher.
  • Expected effect size : a standardised indication of how large the expected result of your study will be, usually based on other similar studies.
  • Population standard deviation: an estimate of the population parameter based on a previous study or a pilot study of your own.

Once you’ve collected all of your data, you can inspect them and calculate descriptive statistics that summarise them.

Inspect your data

There are various ways to inspect your data, including the following:

  • Organising data from each variable in frequency distribution tables .
  • Displaying data from a key variable in a bar chart to view the distribution of responses.
  • Visualising the relationship between two variables using a scatter plot .

By visualising your data in tables and graphs, you can assess whether your data follow a skewed or normal distribution and whether there are any outliers or missing data.

A normal distribution means that your data are symmetrically distributed around a center where most values lie, with the values tapering off at the tail ends.

Mean, median, mode, and standard deviation in a normal distribution

In contrast, a skewed distribution is asymmetric and has more values on one end than the other. The shape of the distribution is important to keep in mind because only some descriptive statistics should be used with skewed distributions.

Extreme outliers can also produce misleading statistics, so you may need a systematic approach to dealing with these values.

Calculate measures of central tendency

Measures of central tendency describe where most of the values in a data set lie. Three main measures of central tendency are often reported:

  • Mode : the most popular response or value in the data set.
  • Median : the value in the exact middle of the data set when ordered from low to high.
  • Mean : the sum of all values divided by the number of values.

However, depending on the shape of the distribution and level of measurement, only one or two of these measures may be appropriate. For example, many demographic characteristics can only be described using the mode or proportions, while a variable like reaction time may not have a mode at all.

Calculate measures of variability

Measures of variability tell you how spread out the values in a data set are. Four main measures of variability are often reported:

  • Range : the highest value minus the lowest value of the data set.
  • Interquartile range : the range of the middle half of the data set.
  • Standard deviation : the average distance between each value in your data set and the mean.
  • Variance : the square of the standard deviation.

Once again, the shape of the distribution and level of measurement should guide your choice of variability statistics. The interquartile range is the best measure for skewed distributions, while standard deviation and variance provide the best information for normal distributions.

Using your table, you should check whether the units of the descriptive statistics are comparable for pretest and posttest scores. For example, are the variance levels similar across the groups? Are there any extreme values? If there are, you may need to identify and remove extreme outliers in your data set or transform your data before performing a statistical test.

From this table, we can see that the mean score increased after the meditation exercise, and the variances of the two scores are comparable. Next, we can perform a statistical test to find out if this improvement in test scores is statistically significant in the population. Example: Descriptive statistics (correlational study) After collecting data from 653 students, you tabulate descriptive statistics for annual parental income and GPA.

It’s important to check whether you have a broad range of data points. If you don’t, your data may be skewed towards some groups more than others (e.g., high academic achievers), and only limited inferences can be made about a relationship.

A number that describes a sample is called a statistic , while a number describing a population is called a parameter . Using inferential statistics , you can make conclusions about population parameters based on sample statistics.

Researchers often use two main methods (simultaneously) to make inferences in statistics.

  • Estimation: calculating population parameters based on sample statistics.
  • Hypothesis testing: a formal process for testing research predictions about the population using samples.

You can make two types of estimates of population parameters from sample statistics:

  • A point estimate : a value that represents your best guess of the exact parameter.
  • An interval estimate : a range of values that represent your best guess of where the parameter lies.

If your aim is to infer and report population characteristics from sample data, it’s best to use both point and interval estimates in your paper.

You can consider a sample statistic a point estimate for the population parameter when you have a representative sample (e.g., in a wide public opinion poll, the proportion of a sample that supports the current government is taken as the population proportion of government supporters).

There’s always error involved in estimation, so you should also provide a confidence interval as an interval estimate to show the variability around a point estimate.

A confidence interval uses the standard error and the z score from the standard normal distribution to convey where you’d generally expect to find the population parameter most of the time.

Hypothesis testing

Using data from a sample, you can test hypotheses about relationships between variables in the population. Hypothesis testing starts with the assumption that the null hypothesis is true in the population, and you use statistical tests to assess whether the null hypothesis can be rejected or not.

Statistical tests determine where your sample data would lie on an expected distribution of sample data if the null hypothesis were true. These tests give two main outputs:

  • A test statistic tells you how much your data differs from the null hypothesis of the test.
  • A p value tells you the likelihood of obtaining your results if the null hypothesis is actually true in the population.

Statistical tests come in three main varieties:

  • Comparison tests assess group differences in outcomes.
  • Regression tests assess cause-and-effect relationships between variables.
  • Correlation tests assess relationships between variables without assuming causation.

Your choice of statistical test depends on your research questions, research design, sampling method, and data characteristics.

Parametric tests

Parametric tests make powerful inferences about the population based on sample data. But to use them, some assumptions must be met, and only some types of variables can be used. If your data violate these assumptions, you can perform appropriate data transformations or use alternative non-parametric tests instead.

A regression models the extent to which changes in a predictor variable results in changes in outcome variable(s).

  • A simple linear regression includes one predictor variable and one outcome variable.
  • A multiple linear regression includes two or more predictor variables and one outcome variable.

Comparison tests usually compare the means of groups. These may be the means of different groups within a sample (e.g., a treatment and control group), the means of one sample group taken at different times (e.g., pretest and posttest scores), or a sample mean and a population mean.

  • A t test is for exactly 1 or 2 groups when the sample is small (30 or less).
  • A z test is for exactly 1 or 2 groups when the sample is large.
  • An ANOVA is for 3 or more groups.

The z and t tests have subtypes based on the number and types of samples and the hypotheses:

  • If you have only one sample that you want to compare to a population mean, use a one-sample test .
  • If you have paired measurements (within-subjects design), use a dependent (paired) samples test .
  • If you have completely separate measurements from two unmatched groups (between-subjects design), use an independent (unpaired) samples test .
  • If you expect a difference between groups in a specific direction, use a one-tailed test .
  • If you don’t have any expectations for the direction of a difference between groups, use a two-tailed test .

The only parametric correlation test is Pearson’s r . The correlation coefficient ( r ) tells you the strength of a linear relationship between two quantitative variables.

However, to test whether the correlation in the sample is strong enough to be important in the population, you also need to perform a significance test of the correlation coefficient, usually a t test, to obtain a p value. This test uses your sample size to calculate how much the correlation coefficient differs from zero in the population.

You use a dependent-samples, one-tailed t test to assess whether the meditation exercise significantly improved math test scores. The test gives you:

  • a t value (test statistic) of 3.00
  • a p value of 0.0028

Although Pearson’s r is a test statistic, it doesn’t tell you anything about how significant the correlation is in the population. You also need to test whether this sample correlation coefficient is large enough to demonstrate a correlation in the population.

A t test can also determine how significantly a correlation coefficient differs from zero based on sample size. Since you expect a positive correlation between parental income and GPA, you use a one-sample, one-tailed t test. The t test gives you:

  • a t value of 3.08
  • a p value of 0.001

The final step of statistical analysis is interpreting your results.

Statistical significance

In hypothesis testing, statistical significance is the main criterion for forming conclusions. You compare your p value to a set significance level (usually 0.05) to decide whether your results are statistically significant or non-significant.

Statistically significant results are considered unlikely to have arisen solely due to chance. There is only a very low chance of such a result occurring if the null hypothesis is true in the population.

This means that you believe the meditation intervention, rather than random factors, directly caused the increase in test scores. Example: Interpret your results (correlational study) You compare your p value of 0.001 to your significance threshold of 0.05. With a p value under this threshold, you can reject the null hypothesis. This indicates a statistically significant correlation between parental income and GPA in male college students.

Note that correlation doesn’t always mean causation, because there are often many underlying factors contributing to a complex variable like GPA. Even if one variable is related to another, this may be because of a third variable influencing both of them, or indirect links between the two variables.

Effect size

A statistically significant result doesn’t necessarily mean that there are important real life applications or clinical outcomes for a finding.

In contrast, the effect size indicates the practical significance of your results. It’s important to report effect sizes along with your inferential statistics for a complete picture of your results. You should also report interval estimates of effect sizes if you’re writing an APA style paper .

With a Cohen’s d of 0.72, there’s medium to high practical significance to your finding that the meditation exercise improved test scores. Example: Effect size (correlational study) To determine the effect size of the correlation coefficient, you compare your Pearson’s r value to Cohen’s effect size criteria.

Decision errors

Type I and Type II errors are mistakes made in research conclusions. A Type I error means rejecting the null hypothesis when it’s actually true, while a Type II error means failing to reject the null hypothesis when it’s false.

You can aim to minimise the risk of these errors by selecting an optimal significance level and ensuring high power . However, there’s a trade-off between the two errors, so a fine balance is necessary.

Frequentist versus Bayesian statistics

Traditionally, frequentist statistics emphasises null hypothesis significance testing and always starts with the assumption of a true null hypothesis.

However, Bayesian statistics has grown in popularity as an alternative approach in the last few decades. In this approach, you use previous research to continually update your hypotheses based on your expectations and observations.

Bayes factor compares the relative strength of evidence for the null versus the alternative hypothesis rather than making a conclusion about rejecting the null hypothesis or not.

Hypothesis testing is a formal procedure for investigating our ideas about the world using statistics. It is used by scientists to test specific predictions, called hypotheses , by calculating how likely it is that a pattern or relationship between variables could have arisen by chance.

The research methods you use depend on the type of data you need to answer your research question .

  • If you want to measure something or test a hypothesis , use quantitative methods . If you want to explore ideas, thoughts, and meanings, use qualitative methods .
  • If you want to analyse a large amount of readily available data, use secondary data. If you want data specific to your purposes with control over how they are generated, collect primary data.
  • If you want to establish cause-and-effect relationships between variables , use experimental methods. If you want to understand the characteristics of a research subject, use descriptive methods.

Statistical analysis is the main method for analyzing quantitative research data . It uses probabilities and models to test predictions about a population from sample data.

Is this article helpful?

Other students also liked, a quick guide to experimental design | 5 steps & examples, controlled experiments | methods & examples of control, between-subjects design | examples, pros & cons, more interesting articles.

  • Central Limit Theorem | Formula, Definition & Examples
  • Central Tendency | Understanding the Mean, Median & Mode
  • Correlation Coefficient | Types, Formulas & Examples
  • Descriptive Statistics | Definitions, Types, Examples
  • How to Calculate Standard Deviation (Guide) | Calculator & Examples
  • How to Calculate Variance | Calculator, Analysis & Examples
  • How to Find Degrees of Freedom | Definition & Formula
  • How to Find Interquartile Range (IQR) | Calculator & Examples
  • How to Find Outliers | Meaning, Formula & Examples
  • How to Find the Geometric Mean | Calculator & Formula
  • How to Find the Mean | Definition, Examples & Calculator
  • How to Find the Median | Definition, Examples & Calculator
  • How to Find the Range of a Data Set | Calculator & Formula
  • Inferential Statistics | An Easy Introduction & Examples
  • Levels of measurement: Nominal, ordinal, interval, ratio
  • Missing Data | Types, Explanation, & Imputation
  • Normal Distribution | Examples, Formulas, & Uses
  • Null and Alternative Hypotheses | Definitions & Examples
  • Poisson Distributions | Definition, Formula & Examples
  • Skewness | Definition, Examples & Formula
  • T-Distribution | What It Is and How To Use It (With Examples)
  • The Standard Normal Distribution | Calculator, Examples & Uses
  • Type I & Type II Errors | Differences, Examples, Visualizations
  • Understanding Confidence Intervals | Easy Examples & Formulas
  • Variability | Calculating Range, IQR, Variance, Standard Deviation
  • What is Effect Size and Why Does It Matter? (Examples)
  • What Is Interval Data? | Examples & Definition
  • What Is Nominal Data? | Examples & Definition
  • What Is Ordinal Data? | Examples & Definition
  • What Is Ratio Data? | Examples & Definition
  • What Is the Mode in Statistics? | Definition, Examples & Calculator

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • Ann Card Anaesth
  • v.22(3); Jul-Sep 2019

Selection of Appropriate Statistical Methods for Data Analysis

Prabhaker mishra.

Department of Biostatistics and Health Informatics, Sanjay Gandhi Post Graduate Institute of Medical Sciences, Lucknow, Uttar Pradesh, India

Chandra Mani Pandey

Uttam singh, amit keshri.

1 Department of Neuro-otology, Sanjay Gandhi Post Graduate Institute of Medical Sciences, Lucknow, Uttar Pradesh, India

Mayilvaganan Sabaretnam

2 Department of Endocrine Surgery, Sanjay Gandhi Post Graduate Institute of Medical Sciences, Lucknow, Uttar Pradesh, India

In biostatistics, for each of the specific situation, statistical methods are available for analysis and interpretation of the data. To select the appropriate statistical method, one need to know the assumption and conditions of the statistical methods, so that proper statistical method can be selected for data analysis. Two main statistical methods are used in data analysis: descriptive statistics, which summarizes data using indexes such as mean and median and another is inferential statistics, which draw conclusions from data using statistical tests such as student's t -test. Selection of appropriate statistical method depends on the following three things: Aim and objective of the study, Type and distribution of the data used, and Nature of the observations (paired/unpaired). All type of statistical methods that are used to compare the means are called parametric while statistical methods used to compare other than means (ex-median/mean ranks/proportions) are called nonparametric methods. In the present article, we have discussed the parametric and non-parametric methods, their assumptions, and how to select appropriate statistical methods for analysis and interpretation of the biomedical data.

Introduction

Selection of appropriate statistical method is very important step in analysis of biomedical data. A wrong selection of the statistical method not only creates some serious problem during the interpretation of the findings but also affects the conclusion of the study. In statistics, for each specific situation, statistical methods are available to analysis and interpretation of the data. To select the appropriate statistical method, one need to know the assumption and conditions of the statistical methods, so that proper statistical method can be selected for data analysis.[ 1 ] Other than knowledge of the statistical methods, another very important aspect is nature and type of the data collected and objective of the study because as per objective, corresponding statistical methods are selected which are suitable on given data. Practice of wrong or inappropriate statistical method is a common phenomenon in the published articles in biomedical research. Incorrect statistical methods can be seen in many conditions like use of unpaired t -test on paired data or use of parametric test for the data which does not follow the normal distribution, etc., At present, many statistical software like SPSS, R, Stata, and SAS are available and using these softwares, one can easily perform the statistical analysis but selection of appropriate statistical test is still a difficult task for the biomedical researchers especially those with nonstatistical background.[ 2 ] Two main statistical methods are used in data analysis: descriptive statistics, which summarizes data using indexes such as mean, median, standard deviation and another is inferential statistics, which draws conclusions from data using statistical tests such as student's t-test, ANOVA test, etc.[ 3 ]

Factors Influencing Selection of Statistical Methods

Selection of appropriate statistical method depends on the following three things: Aim and objective of the study, Type and distribution of the data used, and Nature of the observations (paired/unpaired).

Aim and objective of the study

Selection of statistical test depends upon our aim and objective of the study. Suppose our objective is to find out the predictors of the outcome variable, then regression analysis is used while to compare the means between two independent samples, unpaired samples t-test is used.

Type and distribution of the data used

For the same objective, selection of the statistical test is varying as per data types. For the nominal, ordinal, discrete data, we use nonparametric methods while for continuous data, parametric methods as well as nonparametric methods are used.[ 4 ] For example, in the regression analysis, when our outcome variable is categorical, logistic regression while for the continuous variable, linear regression model is used. The choice of the most appropriate representative measure for continuous variable is dependent on how the values are distributed. If continuous variable follows normal distribution, mean is the representative measure while for non-normal data, median is considered as the most appropriate representative measure of the data set. Similarly in the categorical data, proportion (percentage) while for the ranking/ordinal data, mean ranks are our representative measure. In the inferential statistics, hypothesis is constructed using these measures and further in the hypothesis testing, these measures are used to compare between/among the groups to calculate significance level. Suppose we want to compare the diastolic blood pressure (DBP) between three age groups (years) (<30, 30--50, >50). If our DBP variable is normally distributed, mean value is our representative measure and null hypothesis stated that mean DB P values of the three age groups are statistically equal. In case of non-normal DBP variable, median value is our representative measure and null hypothesis stated that distribution of the DB P values among three age groups are statistically equal. In above example, one-way ANOVA test is used to compare the means when DBP follows normal distribution while Kruskal--Wallis H tests/median tests are used to compare the distribution of DBP among three age groups when DBP follows non-normal distribution. Similarly, suppose we want to compare the mean arterial pressure (MAP) between treatment and control groups, if our MAP variable follows normal distribution, independent samples t-test while in case follow non-normal distribution, Mann--Whitney U test are used to compare the MAP between the treatment and control groups.

Observations are paired or unpaired

Another important point in selection of the statistical test is to assess whether data is paired (same subjects are measures at different time points or using different methods) or unpaired (each group have different subject). For example, to compare the means between two groups, when data is paired, paired samples t-test while for unpaired (independent) data, independent samples t-test is used.

Concept of Parametric and Nonparametric Methods

Inferential statistical methods fall into two possible categorizations: parametric and nonparametric. All type of statistical methods those are used to compare the means are called parametric while statistical methods used to compare other than means (ex-median/mean ranks/proportions) are called nonparametric methods. Parametric tests rely on the assumption that the variable is continuous and follow approximate normally distributed. When data is continuous with non-normal distribution or any other types of data other than continuous variable, nonparametric methods are used. Fortunately, the most frequently used parametric methods have nonparametric counterparts. This can be useful when the assumptions of a parametric test are violated and we can choose the nonparametric alternative as a backup analysis.[ 3 ]

Selection between Parametric and Nonparametric Methods

All type of the t -test, F test are considered parametric test. Student's t -test (one sample t -test, independent samples t -test, paired samples t -test) is used to compare the means between two groups while F test (one-way ANOVA, repeated measures ANOVA, etc.) which is the extension of the student's t -test are used to compare the means among three or more groups. Similarly, Pearson correlation coefficient, linear regression is also considered parametric methods, is used to calculate using mean and standard deviation of the data. For above parametric methods, counterpart nonparametric methods are also available. For example, Mann--Whitney U test and Wilcoxon test are used for student's t -test while Kruskal--Wallis H test, median test, and Friedman test are alternative methods of the F test (ANOVA). Similarly, Spearman rank correlation coefficient and log linear regression are used as nonparametric method of the Pearson correlation and linear regression, respectively.[ 3 , 5 , 6 , 7 , 8 ] Parametric and their counterpart nonparametric methods are given in Table 1 .

Parametric and their Alternative Nonparametric Methods

Statistical Methods to Compare the Proportions

The statistical methods used to compare the proportions are considered nonparametric methods and these methods have no alternative parametric methods. Pearson Chi-square test and Fisher exact test is used to compare the proportions between two or more independent groups. To test the change in proportions between two paired groups, McNemar test is used while Cochran Q test is used for the same objective among three or more paired groups. Z test for proportions is used to compare the proportions between two groups for independent as well as dependent groups.[ 6 , 7 , 8 ] [ Table 2 ].

Other Statistical Methods

Intraclass correlation coefficient is calculated when both pre-post data are in continuous scale. Unweighted and weighted Kappa statistics are used to test the absolute agreement between two methods measured on the same subjects (pre-post) for nominal and ordinal data, respectively. There are some methods those are either semiparametric or nonparametric and these methods, counterpart parametric methods, are not available. Methods are logistic regression analysis, survival analysis, and receiver operating characteristics curve.[ 9 ] Logistic regression analysis is used to predict the categorical outcome variable using independent variable(s). Survival analysis is used to calculate the survival time/survival probability, comparison of the survival time between the groups (Kaplan--Meier method) as well as to identify the predictors of the survival time of the subjects/patients (Cox regression analysis). Receiver operating characteristics (ROC) curve is used to calculate area under curve (AUC) and cutoff values for given continuous variable with corresponding diagnostic accuracy using categorical outcome variable. Diagnostic accuracy of the test method is calculated as compared with another method (usually as compared with gold standard method). Sensitivity (proportion of the detected disease cases from the actual disease cases), specificity (proportion of the detected non-disease subjects from the actual non-disease subjects), overall accuracy (proportion of agreement between test and gold standard methods to correctly detect the disease and non-disease subjects) are the key measures used to assess the diagnostic accuracy of the test method. Other measures like false negative rate (1-sensitivity), false-positive rate (1-specificity), likelihood ratio positive (sensitivity/false-positive rate), likelihood ratio negative (false-negative rate/Specificity), positive predictive value (proportion of correctly detected disease cases by the test variable out of total detected disease cases by the itself), and negative predictive value (proportion of correctly detected non-disease subjects by test variable out of total non-disease subjects detected by the itself) are also used to calculate the diagnostic accuracy of the test method.[ 3 , 6 , 10 ] [ Table 3 ].

Semi-parametric and non-parametric methods

Advantage and Disadvantages of Nonparametric Methods over Parametric Methods and Sample Size Issues

Parametric methods are stronger test to detect the difference between the groups as compared with its counterpart nonparametric methods, although due to some strict assumptions, including normality of the data and sample size, we cannot use parametric test in every situation and resultant its alternative nonparametric methods are used. As mean is used to compare parametric method, which is severally affected by the outliers while in nonparametric method, median/mean rank is our representative measures which do not affect from the outliers.[ 11 ]

In parametric methods like student's t-test and ANOVA test, significance level is calculated using mean and standard deviation, and to calculate standard deviation in each group, at least two observations are required. If every group did not have at least two observations, its alternative nonparametric method to be selected works through comparisons of the mean ranks of the data.

For small sample size (average ≤15 observations per group), normality testing methods are less sensitive about non-normality and there is chance to detect normality despite having non-normal data. It is recommended that when sample size is small, only on highly normally distributed data, parametric method should be used otherwise corresponding nonparametric methods should be preferred. Similarly on sufficient or large sample size (average >15 observations per group), most of the statistical methods are highly sensitive about non-normality and there is chance to wrongly detect non-normality, despite having normal data. It is recommended that when sample size is sufficient, only on highly non-normal data, nonparametric method should be used otherwise corresponding parametric methods should be preferred.[ 12 ]

Minimum Sample Size Required for Statistical Methods

To detect the significant difference between the means/medians/mean ranks/proportions, at minimum level of confidence (usually 95%) and power of the test (usually 80%), how many individuals/subjects (sample size) are required depends on the detected effect size. The effect size and corresponding required sample size are inversely proportional to each other, that is, on the same level of confidence and power of the test, when effect size is increasing, required sample size is decreasing. Summary is, no minimum or maximum sample size is fix for any particular statistical method and it is subject to estimate based on the given inputs including effect size, level of confidence, power of the study, etc., Only on the sufficient sample size, we can detect the difference significantly. In case lack of the sample size than actual required, our study will be under power to detect the given difference as well as result would be statistically insignificant.

Impact of Wrong Selection of the Statistical Methods

As for each and every situation, there are specific statistical methods. Failing to select appropriate statistical method, our significance level as well as their conclusion is affected.[ 13 ] For example in a study, systolic blood pressure (mean ± SD) of the control (126.45 ± 8.85, n 1 =20) and treatment (121.85 ± 5.96, n 2 =20) group was compared using Independent samples t -test (correct practice). Result showed that mean difference between two groups was statistically insignificant ( P = 0.061) while on the same data, paired samples t -test (incorrect practice) indicated that mean difference was statistically significant ( P = 0.011). Due to incorrect practice, we detected the statistically significant difference between the groups although actually difference did not exist.

Conclusions

Selection of the appropriate statistical methods is very important for the quality research. It is important that a researcher knows the basic concepts of the statistical methods used to conduct research study that produce a valid and reliable results. There are various statistical methods that can be used in different situations. Each test makes particular assumptions about the data. These assumptions should be taken into consideration when deciding which the most appropriate test is. Wrong or inappropriate use of statistical methods may lead to defective conclusions, finally would harm the evidence-based practices. Hence, an adequate knowledge of statistics and the appropriate use of statistical tests are important for improving and producing quality biomedical research. However, it is extremely difficult for a biomedical researchers or academician to learn the entire statistical methods. Therefore, at least basic knowledge is very important so that appropriate selection of the statistical methods can decide as well as correct/incorrect practices can be recognized in the published research. There are many softwares available online as well as offline for analyzing the data, although it is fact that which set of statistical tests are appropriate for the given data and study objective is still very difficult for the researchers to understand. Therefore, since planning of the study to data collection, analysis and finally in the review process, proper consultation from statistical experts may be an alternative option and can reduce the burden from the clinicians to go in depth of statistics which required lots of time and effort and ultimately affect their clinical works. These practices not only ensure the correct and appropriate use of the biostatistical methods in the research but also ensure the highest quality of statistical reporting in the research and journals.[ 14 ]

Financial support and sponsorship

Conflicts of interest.

There are no conflicts of interest.

Acknowledgements

Authors would like to express their deep and sincere gratitude to Dr. Prabhat Tiwari, Professor, Department of Anaesthesiology, Sanjay Gandhi Postgraduate Institute of Medical Sciences, Lucknow, for his encouragement to write this article. His critical reviews and suggestions were very useful for improvement in the article.

StatAnalytica

How To Write a Statistical Research Paper: Tips, Topics, Outline

Statistical Research Paper

Working on a research paper can be a bit challenging. Some people even opt for paying online writing companies to do the job for them. While this might seem like a better solution, it can cost you a lot of money. A cheaper option is to search online for the critical parts of your essay. Your data should come from reliable sources for your research paper to be authentic. You will also need to introduce your work to your readers. It should be straightforward and relevant to the topic.  With this in mind, here is a guideline to help you succeed in your research writing. But before that, let’s see what the outline should look like.

The Outline

Table of Contents

How to write a statistical analysis paper is a puzzle many people find difficult to crack. It’s not such a challenging task as you might think, especially if you learn some helpful tips to make the writing process easier. It’s just like working on any other essay. You only need to get the format and structure right and study the process. Here is what the general outline should look like:

  • introduction;
  • problem statement;
  • objectives;
  • methodology;
  • data examination;
  • discussion;
  • conclusion and recommendations.

Let us now see some tips that can help you become a better statistical researcher.

  • Top 99+ Trending Statistics Research Topics for Students

Tips for Writing Statistics Research Paper

If you are wondering how people write their papers, you are in the right place. We’ll take a look at a few pointers that can help you come up with amazing work.

Choose A Topic

Basically, this is the most important stage of your essay. Whether you want to pay for it or not, you need a simple and accessible topic to write about. Usually, the paid research papers have a well-formed and clear topic. It helps your paper to stand out. Start off by explaining to your audience what your papers are all about. Also, check whether there is enough data to support your idea. The weaker the topic is, the harder your work will be. Is the potential theme within the realm of statistics? Can the question at hand be solved with the help of the available data? These are some of the questions someone should answer first. In the end, the topic you opt for should provide sufficient space for independent information collection and analysis.

Collect Data

This stage relies heavily on the quantity of data sources and the method used to collect them. Keep in mind that you must stick to the chosen methodology throughout your essay. It is also important to explain why you opted for the data collection method used. Plus, be cautious when collecting information. One simple mistake can compromise the entire work. You can source your data from reliable sources like google, read published articles, or experiment with your own findings. However, if your instructor provides certain recommendations, follow them instead. Don’t twist the information to fit your interest to avoid losing originality. And in case no recommendations are given, ask your instructor to provide some.

Write Body Paragraphs

Use the information garnered to create the main body of your essay. After identifying an applicable area of interest, use the data to build your paragraphs. You can start off by making a rough draft of your findings and then use it as a guide for your main essay. The next step is to construe numerical figures and make conclusions. This stage requires your proficiency in interpreting statistics. Integrate your math engagement strategies to break down those figures and pinpoint only the most meaningful parts of them. Also, include some common counterpoints and support the information with specific examples.

Create Your Essay

Now that you have all the appropriate materials at hand, this section will be easy. Simply note down all the information gathered, citing your sources as well. Make sure not to copy and paste directly to avoid plagiarism. Your content should be unique and easy to read, too. We recommend proofreading and polishing your work before making it public. In addition, be on the lookout for any grammatical, spelling, or punctuation mistakes.

This section is a summary of all your findings. Explain the importance of what you are doing. You can also include suggestions for future work. Make sure to restate what you mentioned in the introduction and touch a little bit on the method used to collect and analyze your data. In short, sum up everything you’ve written in your essay.

How to Find Statistical Topics for your Paper

Statistics is a discipline that involves collecting, analyzing, organizing, presenting, and interpreting data. If you are looking for the right topic for your work, here are a few things to consider.

●   Start by finding out what topics have already been worked on and pick the remaining areas.

●   Consider recent developments in your field of study that may inspire a new topic.

●   Think about any specific questions or problems that you have come across on your own that could be explored further.

●   Ask your advisor or mentor for suggestions.

●   Review conference proceedings, journal articles, and other publications.

●   Try using a brainstorming technique. For instance, list out related keywords and combine them in different ways to generate new ideas.

Try out some of these tips. Be sure to find something that will work for you.

Working on a statistics paper can be quite challenging to work on. But with the right information sources, everything becomes easy. This guide will help you reveal the secret of preparing such essays. Also, don’t forget to do more reading to broaden your knowledge. You can find statistics research paper examples and refer to them for ideas. Nonetheless, if you’re still not confident enough, you can always hire a trustworthy writing company to get the job done.

Related Posts

application-of-data-miningapplication-of-data-mining

Top Useful Applications of Data Mining in Different Fields

WHAT ARE THE BENEFICIAL FACTORS OF RENTING 1 BHK FLATS IN HYDERABAD?

What Are The Beneficial Factors Of Renting 1 BHK Flats In Hyderabad?

Statistics Research Paper

Academic Writing Service

View sample Statistics Research Paper. Browse other  research paper examples and check the list of research paper topics for more inspiration. If you need a religion research paper written according to all the academic standards, you can always turn to our experienced writers for help. This is how your paper can get an A! Feel free to contact our custom writing service s for professional assistance. We offer high-quality assignments for reasonable rates.

Statistics Research Paper

Academic Writing, Editing, Proofreading, And Problem Solving Services

Get 10% off with 24start discount code, more statistics research papers:.

  • Time Series Research Paper
  • Crime Statistics Research Paper
  • Economic Statistics Research Paper
  • Education Statistics Research Paper
  • Health Statistics Research Paper
  • Labor Statistics Research Paper
  • History of Statistics Research Paper
  • Survey Sampling Research Paper
  • Multidimensional Scaling Research Paper
  • Sequential Statistical Methods Research Paper
  • Simultaneous Equation Estimation Research Paper
  • Statistical Clustering Research Paper
  • Statistical Sufficiency Research Paper
  • Censuses Of Population Research Paper
  • Stochastic Models Research Paper
  • Stock Market Predictability Research Paper
  • Structural Equation Modeling Research Paper
  • Survival Analysis Research Paper
  • Systems Modeling Research Paper
  • Nonprobability Sampling Research Paper

1. Introduction

Statistics is a body of quantitative methods associated with empirical observation. A primary goal of these methods is coping with uncertainty. Most formal statistical methods rely on probability theory to express this uncertainty and to provide a formal mathematical basis for data description and for analysis. The notion of variability associated with data, expressed through probability, plays a fundamental role in this theory. As a consequence, much statistical effort is focused on how to control and measure variability and/or how to assign it to its sources.

Almost all characterizations of statistics as a field include the following elements:

(a) Designing experiments, surveys, and other systematic forms of empirical study.

(b) Summarizing and extracting information from data.

(c) Drawing formal inferences from empirical data through the use of probability.

(d) Communicating the results of statistical investigations to others, including scientists, policy makers, and the public.

This research paper describes a number of these elements, and the historical context out of which they grew. It provides a broad overview of the field, that can serve as a starting point to many of the other statistical entries in this encyclopedia.

2. The Origins Of The Field of Statistics

The word ‘statistics’ is related to the word ‘state’ and the original activity that was labeled as statistics was social in nature and related to elements of society through the organization of economic, demographic, and political facts. Paralleling this work to some extent was the development of the probability calculus and the theory of errors, typically associated with the physical sciences. These traditions came together in the nineteenth century and led to the notion of statistics as a collection of methods for the analysis of scientific data and the drawing of inferences therefrom.

As Hacking (1990) has noted: ‘By the end of the century chance had attained the respectability of a Victorian valet, ready to be the logical servant of the natural, biological and social sciences’ ( p. 2). At the beginning of the twentieth century, we see the emergence of statistics as a field under the leadership of Karl Pearson, George Udny Yule, Francis Y. Edgeworth, and others of the ‘English’ statistical school. As Stigler (1986) suggests:

Before 1900 we see many scientists of different fields developing and using techniques we now recognize as belonging to modern statistics. After 1900 we begin to see identifiable statisticians developing such techniques into a unified logic of empirical science that goes far beyond its component parts. There was no sharp moment of birth; but with Pearson and Yule and the growing number of students in Pearson’s laboratory, the infant discipline may be said to have arrived. (p. 361)

Pearson’s laboratory at University College, London quickly became the first statistics department in the world and it was to influence subsequent developments in a profound fashion for the next three decades. Pearson and his colleagues founded the first methodologically-oriented statistics journal, Biometrika, and they stimulated the development of new approaches to statistical methods. What remained before statistics could legitimately take on the mantle of a field of inquiry, separate from mathematics or the use of statistical approaches in other fields, was the development of the formal foundations of theories of inference from observations, rooted in an axiomatic theory of probability.

Beginning at least with the Rev. Thomas Bayes and Pierre Simon Laplace in the eighteenth century, most early efforts at statistical inference used what was known as the method of inverse probability to update a prior probability using the observed data in what we now refer to as Bayes’ Theorem. (For a discussion of who really invented Bayes’ Theorem, see Stigler 1999, Chap. 15). Inverse probability came under challenge in the nineteenth century, but viable alternative approaches gained little currency. It was only with the work of R. A. Fisher on statistical models, estimation, and significance tests, and Jerzy Neyman and Egon Pearson, in the 1920s and 1930s, on tests of hypotheses, that alternative approaches were fully articulated and given a formal foundation. Neyman’s advocacy of the role of probability in the structuring of a frequency-based approach to sample surveys in 1934 and his development of confidence intervals further consolidated this effort at the development of a foundation for inference (cf. Statistical Methods, History of: Post- 1900 and the discussion of ‘The inference experts’ in Gigerenzer et al. 1989).

At about the same time Kolmogorov presented his famous axiomatic treatment of probability, and thus by the end of the 1930s, all of the requisite elements were finally in place for the identification of statistics as a field. Not coincidentally, the first statistical society devoted to the mathematical underpinnings of the field, The Institute of Mathematical Statistics, was created in the United States in the mid-1930s. It was during this same period that departments of statistics and statistical laboratories and groups were first formed in universities in the United States.

3. Emergence Of Statistics As A Field

3.1 the role of world war ii.

Perhaps the greatest catalysts to the emergence of statistics as a field were two major social events: the Great Depression of the 1930s and World War II. In the United States, one of the responses to the depression was the development of large-scale probability-based surveys to measure employment and unemployment. This was followed by the institutionalization of sampling as part of the 1940 US decennial census. But with World War II raging in Europe and in Asia, mathematicians and statisticians were drawn into the war effort, and as a consequence they turned their attention to a broad array of new problems. In particular, multiple statistical groups were established in both England and the US specifically to develop new methods and to provide consulting. (See Wallis 1980, on statistical groups in the US; Barnard and Plackett 1985, for related efforts in the United Kingdom; and Fienberg 1985). These groups not only created imaginative new techniques such as sequential analysis and statistical decision theory, but they also developed a shared research agenda. That agenda led to a blossoming of statistics after the war, and in the 1950s and 1960s to the creation of departments of statistics at universities—from coast to coast in the US, and to a lesser extent in England and elsewhere.

3.2 The Neo-Bayesian Revival

Although inverse probability came under challenge in the 1920s and 1930s, it was not totally abandoned. John Maynard Keynes (1921) wrote A Treatise on Probability that was rooted in this tradition, and Frank Ramsey (1926) provided an early effort at justifying the subjective nature of prior distributions and suggested the importance of utility functions as an adjunct to statistical inference. Bruno de Finetti provided further development of these ideas in the 1930s, while Harold Jeffreys (1938) created a separate ‘objective’ development of these and other statistical ideas on inverse probability.

Yet as statistics flourished in the post-World War II era, it was largely based on the developments of Fisher, Neyman and Pearson, as well as the decision theory methods of Abraham Wald (1950). L. J. Savage revived interest in the inverse probability approach with The Foundations of Statistics (1954) in which he attempted to provide the axiomatic foundation from the subjective perspective. In an essentially independent effort, Raiffa and Schlaifer (1961) attempted to provide inverse probability counterparts to many of the then existing frequentist tools, referring to these alternatives as ‘Bayesian.’ By 1960, the term ‘Bayesian inference’ had become standard usage in the statistical literature, the theoretical interest in the development of Bayesian approaches began to take hold, and the neo-Bayesian revival was underway. But the movement from Bayesian theory to statistical practice was slow, in large part because the computations associated with posterior distributions were an overwhelming stumbling block for those who were interested in the methods. Only in the 1980s and 1990s did new computational approaches revolutionize both Bayesian methods, and the interest in them, in a broad array of areas of application.

3.3 The Role Of Computation In Statistics

From the days of Pearson and Fisher, computation played a crucial role in the development and application of statistics. Pearson’s laboratory employed dozens of women who used mechanical devices to carry out the careful and painstaking calculations required to tabulate values from various probability distributions. This effort ultimately led to the creation of the Biometrika Tables for Statisticians that were so widely used by others applying tools such as chisquare tests and the like. Similarly, Fisher also developed his own set of statistical tables with Frank Yates when he worked at Rothamsted Experiment Station in the 1920s and 1930s. One of the most famous pictures of Fisher shows him seated at Whittingehame Lodge, working at his desk calculator (see Box 1978).

The development of the modern computer revolutionized statistical calculation and practice, beginning with the creation of the first statistical packages in the 1960s—such as the BMDP package for biological and medical applications, and Datatext for statistical work in the social sciences. Other packages soon followed—such as SAS and SPSS for both data management and production-like statistical analyses, and MINITAB for the teaching of statistics. In 2001, in the era of the desktop personal computer, almost everyone has easy access to interactive statistical programs that can implement complex statistical procedures and produce publication-quality graphics. And there is a new generation of statistical tools that rely upon statistical simulation such as the bootstrap and Markov Chain Monte Carlo methods. Complementing the traditional production-like packages for statistical analysis are more methodologically oriented languages such as S and S-PLUS, and symbolic and algebraic calculation packages. Statistical journals and those in various fields of application devote considerable space to descriptions of such tools.

4. Statistics At The End Of The Twentieth Century

It is widely recognized that any statistical analysis can only be as good as the underlying data. Consequently, statisticians take great care in the the design of methods for data collection and in their actual implementation. Some of the most important modes of statistical data collection include censuses, experiments, observational studies, and sample Surveys, all of which are discussed elsewhere in this encyclopedia. Statistical experiments gain their strength and validity both through the random assignment of treatments to units and through the control of nontreatment variables. Similarly sample surveys gain their validity for generalization through the careful design of survey questionnaires and probability methods used for the selection of the sample units. Approaches to cope with the failure to fully implement randomization in experiments or random selection in sample surveys are discussed in Experimental Design: Compliance and Nonsampling Errors.

Data in some statistical studies are collected essentially at a single point in time (cross-sectional studies), while in others they are collected repeatedly at several time points or even continuously, while in yet others observations are collected sequentially, until sufficient information is available for inferential purposes. Different entries discuss these options and their strengths and weaknesses.

After a century of formal development, statistics as a field has developed a number of different approaches that rely on probability theory as a mathematical basis for description, analysis, and statistical inference. We provide an overview of some of these in the remainder of this section and provide some links to other entries in this encyclopedia.

4.1 Data Analysis

The least formal approach to inference is often the first employed. Its name stems from a famous article by John Tukey (1962), but it is rooted in the more traditional forms of descriptive statistical methods used for centuries.

Today, data analysis relies heavily on graphical methods and there are different traditions, such as those associated with

(a) The ‘exploratory data analysis’ methods suggested by Tukey and others.

(b) The more stylized correspondence analysis techniques of Benzecri and the French school.

(c) The alphabet soup of computer-based multivariate methods that have emerged over the past decade such as ACE, MARS, CART, etc.

No matter which ‘school’ of data analysis someone adheres to, the spirit of the methods is typically to encourage the data to ‘speak for themselves.’ While no theory of data analysis has emerged, and perhaps none is to be expected, the flexibility of thought and method embodied in the data analytic ideas have influenced all of the other approaches.

4.2 Frequentism

The name of this group of methods refers to a hypothetical infinite sequence of data sets generated as was the data set in question. Inferences are to be made with respect to this hypothetical infinite sequence. (For details, see Frequentist Inference).

One of the leading frequentist methods is significance testing, formalized initially by R. A. Fisher (1925) and subsequently elaborated upon and extended by Neyman and Pearson and others (see below). Here a null hypothesis is chosen, for example, that the mean, µ, of a normally distributed set of observations is 0. Fisher suggested the choice of a test statistic, e.g., based on the sample mean, x, and the calculation of the likelihood of observing an outcome as or more extreme as x is from µ 0, a quantity usually labeled as the p-value. When p is small (e.g., less than 5 percent), either a rare event has occurred or the null hypothesis is false. Within this theory, no probability can be given for which of these two conclusions is the case.

A related set of methods is testing hypotheses, as proposed by Neyman and Pearson (1928, 1932). In this approach, procedures are sought having the property that, for an infinite sequence of such sets, in only (say) 5 percent for would the null hypothesis be rejected if the null hypothesis were true. Often the infinite sequence is restricted to sets having the same sample size, but this is unnecessary. Here, in addition to the null hypothesis, an alternative hypothesis is specified. This permits the definition of a power curve, reflecting the frequency of rejecting the null hypothesis when the specified alternative is the case. But, as with the Fisherian approach, no probability can be given to either the null or the alternative hypotheses.

The construction of confidence intervals, following the proposal of Neyman (1934), is intimately related to testing hypotheses; indeed a 95 percent confidence interval may be regarded as the set of null hypotheses which, had they been tested at the 5 percent level of significance, would not have been rejected. A confidence interval is a random interval, having the property that the specified proportion (say 95 percent) of the infinite sequence, of random intervals would have covered the true value. For example, an interval that 95 percent of the time (by auxiliary randomization) is the whole real line, and 5 percent of the time is the empty set, is a valid 95 percent confidence interval.

Estimation of parameters—i.e., choosing a single value of the parameters that is in some sense best—is also an important frequentist method. Many methods have been proposed, both for particular models and as general approaches regardless of model, and their frequentist properties explored. These methods usually extended to intervals of values through inversion of test statistics or via other related devices. The resulting confidence intervals share many of the frequentist theoretical properties of the corresponding test procedures.

Frequentist statisticians have explored a number of general properties thought to be desirable in a procedure, such as invariance, unbiasedness, sufficiency, conditioning on ancillary statistics, etc. While each of these properties has examples in which it appears to produce satisfactory recommendations, there are others in which it does not. Additionally, these properties can conflict with each other. No general frequentist theory has emerged that proposes a hierarchy of desirable properties, leaving a frequentist without guidance in facing a new problem.

4.3 Likelihood Methods

The likelihood function (first studied systematically by R. A. Fisher) is the probability density of the data, viewed as a function of the parameters. It occupies an interesting middle ground in the philosophical debate, as it is used both by frequentists (as in maximum likelihood estimation) and by Bayesians in the transition from prior distributions to posterior distributions. A small group of scholars (among them G. A. Barnard, A. W. F. Edwards, R. Royall, D. Sprott) have proposed the likelihood function as an independent basis for inference. The issue of nuisance parameters has perplexed this group, since maximization, as would be consistent with maximum likelihood estimation, leads to different results in general than does integration, which would be consistent with Bayesian ideas.

4.4 Bayesian Methods

Both frequentists and Bayesians accept Bayes’ Theorem as correct, but Bayesians use it far more heavily. Bayesian analysis proceeds from the idea that probability is personal or subjective, reflecting the views of a particular person at a particular point in time. These views are summarized in the prior distribution over the parameter space. Together the prior distribution and the likelihood function define the joint distribution of the parameters and the data. This joint distribution can alternatively be factored as the product of the posterior distribution of the parameter given the data times the predictive distribution of the data.

In the past, Bayesian methods were deemed to be controversial because of the avowedly subjective nature of the prior distribution. But the controversy surrounding their use has lessened as recognition of the subjective nature of the likelihood has spread. Unlike frequentist methods, Bayesian methods are, in principle, free of the paradoxes and counterexamples that make classical statistics so perplexing. The development of hierarchical modeling and Markov Chain Monte Carlo (MCMC) methods have further added to the current popularity of the Bayesian approach, as they allow analyses of models that would otherwise be intractable.

Bayesian decision theory, which interacts closely with Bayesian statistical methods, is a useful way of modeling and addressing decision problems of experimental designs and data analysis and inference. It introduces the notion of utilities and the optimum decision combines probabilities of events with utilities by the calculation of expected utility and maximizing the latter (e.g., see the discussion in Lindley 2000).

Current research is attempting to use the Bayesian approach to hypothesis testing to provide tests and pvalues with good frequentist properties (see Bayarri and Berger 2000).

4.5 Broad Models: Nonparametrics And Semiparametrics

These models include parameter spaces of infinite dimensions, whether addressed in a frequentist or Bayesian manner. In a sense, these models put more inferential weight on the assumption of conditional independence than does an ordinary parametric model.

4.6 Some Cross-Cutting Themes

Often different fields of application of statistics need to address similar issues. For example, dimensionality of the parameter space is often a problem. As more parameters are added, the model will in general fit better (at least no worse). Is the apparent gain in accuracy worth the reduction in parsimony? There are many different ways to address this question in the various applied areas of statistics.

Another common theme, in some sense the obverse of the previous one, is the question of model selection and goodness of fit. In what sense can one say that a set of observations is well-approximated by a particular distribution? (cf. Goodness of Fit: Overview). All statistical theory relies at some level on the use of formal models, and the appropriateness of those models and their detailed specification are of concern to users of statistical methods, no matter which school of statistical inference they choose to work within.

5. Statistics In The Twenty-first Century

5.1 adapting and generalizing methodology.

Statistics as a field provides scientists with the basis for dealing with uncertainty, and, among other things, for generalizing from a sample to a population. There is a parallel sense in which statistics provides a basis for generalization: when similar tools are developed within specific substantive fields, such as experimental design methodology in agriculture and medicine, and sample surveys in economics and sociology. Statisticians have long recognized the common elements of such methodologies and have sought to develop generalized tools and theories to deal with these separate approaches (see e.g., Fienberg and Tanur 1989).

One hallmark of modern statistical science is the development of general frameworks that unify methodology. Thus the tools of Generalized Linear Models draw together methods for linear regression and analysis of various models with normal errors and those log-linear and logistic models for categorical data, in a broader and richer framework. Similarly, graphical models developed in the 1970s and 1980s use concepts of independence to integrate work in covariance section, decomposable log-linear models, and Markov random field models, and produce new methodology as a consequence. And the latent variable approaches from psychometrics and sociology have been tied with simultaneous equation and measurement error models from econometrics into a broader theory of covariance analysis and structural equations models.

Another hallmark of modern statistical science is the borrowing of methods in one field for application in another. One example is provided by Markov Chain Monte Carlo methods, now used widely in Bayesian statistics, which were first used in physics. Survival analysis, used in biostatistics to model the disease-free time or time-to-mortality of medical patients, and analyzed as reliability in quality control studies, are now used in econometrics to measure the time until an unemployed person gets a job. We anticipate that this trend of methodological borrowing will continue across fields of application.

5.2 Where Will New Statistical Developments Be Focused ?

In the issues of its year 2000 volume, the Journal of the American Statistical Association explored both the state of the art of statistics in diverse areas of application, and that of theory and methods, through a series of vignettes or short articles. These essays provide an excellent supplement to the entries of this encyclopedia on a wide range of topics, not only presenting a snapshot of the current state of play in selected areas of the field but also affecting some speculation on the next generation of developments. In an afterword to the last set of these vignettes, Casella (2000) summarizes five overarching themes that he observed in reading through the entire collection:

(a) Large datasets.

(b) High-dimensional/nonparametric models.

(c) Accessible computing.

(d) Bayes/frequentist/who cares?

(e) Theory/applied/why differentiate?

Not surprisingly, these themes fit well those that one can read into the statistical entries in this encyclopedia. The coming together of Bayesian and frequentist methods, for example, is illustrated by the movement of frequentists towards the use of hierarchical models and the regular consideration of frequentist properties of Bayesian procedures (e.g., Bayarri and Berger 2000). Similarly, MCMC methods are being widely used in non-Bayesian settings and, because they focus on long-run sequences of dependent draws from multivariate probability distributions, there are frequentist elements that are brought to bear in the study of the convergence of MCMC procedures. Thus the oft-made distinction between the different schools of statistical inference (suggested in the preceding section) is not always clear in the context of real applications.

5.3 The Growing Importance Of Statistics Across The Social And Behavioral Sciences

Statistics touches on an increasing number of fields of application, in the social sciences as in other areas of scholarship. Historically, the closest links have been with economics; together these fields share parentage of econometrics. There are now vigorous interactions with political science, law, sociology, psychology, anthropology, archeology, history, and many others.

In some fields, the development of statistical methods has not been universally welcomed. Using these methods well and knowledgeably requires an understanding both of the substantive field and of statistical methods. Sometimes this combination of skills has been difficult to develop.

Statistical methods are having increasing success in addressing questions throughout the social and behavioral sciences. Data are being collected and analyzed on an increasing variety of subjects, and the analyses are becoming increasingly sharply focused on the issues of interest.

We do not anticipate, nor would we find desirable, a future in which only statistical evidence was accepted in the social and behavioral sciences. There is room for, and need for, many different approaches. Nonetheless, we expect the excellent progress made in statistical methods in the social and behavioral sciences in recent decades to continue and intensify.

Bibliography:

  • Barnard G A, Plackett R L 1985 Statistics in the United Kingdom, 1939–1945. In: Atkinson A C, Fienberg S E (eds.) A Celebration of Statistics: The ISI Centennial Volume. Springer-Verlag, New York, pp. 31–55
  • Bayarri M J, Berger J O 2000 P values for composite null models (with discussion). Journal of the American Statistical Association 95: 1127–72
  • Box J 1978 R. A. Fisher, The Life of a Scientist. Wiley, New York
  • Casella G 2000 Afterword. Journal of the American Statistical Association 95: 1388
  • Fienberg S E 1985 Statistical developments in World War II: An international perspective. In: Anthony C, Atkinson A C, Fienberg S E (eds.) A Celebration of Statistics: The ISI Centennial Volume. Springer-Verlag, New York, pp. 25–30
  • Fienberg S E, Tanur J M 1989 Combining cognitive and statistical approaches to survey design. Science 243: 1017–22
  • Fisher R A 1925 Statistical Methods for Research Workers. Oliver and Boyd, London
  • Gigerenzer G, Swijtink Z, Porter T, Daston L, Beatty J, Kruger L 1989 The Empire of Chance. Cambridge University Press, Cambridge, UK
  • Hacking I 1990 The Taming of Chance. Cambridge University Press, Cambridge, UK
  • Jeffreys H 1938 Theory of Probability, 2nd edn. Clarendon Press, Oxford, UK
  • Keynes J 1921 A Treatise on Probability. Macmillan, London
  • Lindley D V 2000/1932 The philosophy of statistics (with discussion). The Statistician 49: 293–337
  • Neyman J 1934 On the two different aspects of the representative method: the method of stratified sampling and the method of purposive selection (with discussion). Journal of the Royal Statistical Society 97: 558–625
  • Neyman J, Pearson E S 1928 On the use and interpretation of certain test criteria for purposes of statistical inference. Part I. Biometrika 20A: 175–240
  • Neyman J, Pearson E S 1932 On the problem of the most efficient tests of statistical hypotheses. Philosophical Transactions of the Royal Society, Series. A 231: 289–337
  • Raiffa H, Schlaifer R 1961 Applied Statistical Decision Theory. Harvard Business School, Boston
  • Ramsey F P 1926 Truth and probability. In: The Foundations of Mathematics and Other Logical Essays. Kegan Paul, London, pp.
  • Savage L J 1954 The Foundations of Statistics. Wiley, New York
  • Stigler S M 1986 The History of Statistics: The Measurement of Uncertainty Before 1900. Harvard University Press, Cambridge, MA
  • Stigler S M 1999 Statistics on the Table: The History of Statistical Concepts and Methods. Harvard University Press, Cambridge, MA
  • Tukey John W 1962 The future of data analysis. Annals of Mathematical Statistics 33: 1–67
  • Wald A 1950 Statistical Decision Functions. Wiley, New York
  • Wallis W 1980 The Statistical Research Group, 1942–1945 (with discussion). Journal of the American Statistical Association 75: 320–35

ORDER HIGH QUALITY CUSTOM PAPER

example of statistical analysis research paper

Statistical Research Questions: Five Examples for Quantitative Analysis

Table of contents, introduction.

How are statistical research questions for quantitative analysis written? This article provides five examples of statistical research questions that will allow statistical analysis to take place.

In quantitative research projects, writing statistical research questions requires a good understanding and the ability to discern the type of data that you will analyze. This knowledge is elemental in framing research questions that shall guide you in identifying the appropriate statistical test to use in your research.

Thus, before writing your statistical research questions and reading the examples in this article, read first the article that enumerates the  four types of measurement scales . Knowing the four types of measurement scales will enable you to appreciate the formulation or structuring of research questions.

Once you feel confident that you can correctly identify the nature of your data, the following examples of statistical research questions will strengthen your understanding. Asking these questions can help you unravel unexpected outcomes or discoveries particularly while doing exploratory data analysis .

Five Examples of Statistical Research Questions

In writing the statistical research questions, I provide a topic that shows the variables of the study, the study description, and a link to the original scientific article to give you a glimpse of the real-world examples.

Topic 1: Physical Fitness and Academic Achievement

A study was conducted to determine the relationship between physical fitness and academic achievement. The subjects of the study include school children in urban schools.

Statistical Research Question No. 1

Is there a significant relationship between physical fitness and academic achievement?

Notice that this study correlated two variables, namely 1) physical fitness, and 2) academic achievement.

To allow statistical analysis to take place, there is a need to define what is physical fitness, as well as academic achievement. The researchers measured physical fitness in terms of  the number of physical fitness tests  that the students passed during their physical education class. It’s simply counting the ‘number of PE tests passed.’

On the other hand, the researchers measured academic achievement in terms of a passing score in Mathematics and English. The variable is the  number of passing scores  in both Mathematics and English.

Both variables are ratio variables. 

Given the statistical research question, the appropriate statistical test can be applied to determine the relationship. A Pearson correlation coefficient test will test the significance and degree of the relationship. But the more sophisticated higher level statistical test can be applied if there is a need to correlate with other variables.

In the particular study mentioned, the researchers used  multivariate logistic regression analyses  to assess the probability of passing the tests, controlling for students’ weight status, ethnicity, gender, grade, and socioeconomic status. For the novice researcher, this requires further study of multivariate (or many variables) statistical tests. You may study it on your own.

Most of what I discuss in the statistics articles I wrote came from self-study. It’s easier to understand concepts now as there are a lot of resource materials available online. Videos and ebooks from places like Youtube, Veoh, The Internet Archives, among others, provide free educational materials. Online education will be the norm of the future. I describe this situation in my post about  Education 4.0 .

The following video sheds light on the frequently used statistical tests and their selection. It is an excellent resource for beginners. Just maintain an open mind to get rid of your dislike for numbers; that is, if you are one of those who have a hard time understanding mathematical concepts. My ebook on  statistical tests and their selection  provides many examples.

Source: Chomitz et al. (2009)

Topic 2: Climate Conditions and Consumption of Bottled Water

This study attempted to correlate climate conditions with the decision of people in Ecuador to consume bottled water, including the volume consumed. Specifically, the researchers investigated if the increase in average ambient temperature affects the consumption of bottled water.

Statistical Research Question No. 2

Is there a significant relationship between average temperature and amount of bottled water consumed?

In this instance, the variables measured include the  average temperature in the areas studied  and the  volume of water consumed . Temperature is an  interval variable,  while volume is a  ratio variable .

In this example, the variables include the  average temperature  and  volume of bottled water . The first variable (average temperature) is an interval variable, and the latter (volume of water) is a ratio variable.

Now, it’s easy to identify the statistical test to analyze the relationship between the two variables. You may refer to my previous post titled  Parametric Statistics: Four Widely Used Parametric Tests and When to Use Them . Using the figure supplied in that article, the appropriate test to use is, again, Pearson’s Correlation Coefficient.

Source: Zapata (2021)

Topic 3: Nursing Home Staff Size and Number of COVID-19 Cases

research question

An investigation sought to determine if the size of nursing home staff and the number of COVID-19 cases are correlated. Specifically, they looked into the number of unique employees working daily, and the outcomes include weekly counts of confirmed COVID-19 cases among residents and staff and weekly COVID-19 deaths among residents.

Statistical Research Question No. 3

Is there a significant relationship between the number of unique employees working in skilled nursing homes and the following:

  • number of weekly confirmed COVID-19 cases among residents and staff, and
  • number of weekly COVID-19 deaths among residents.

Note that this study on COVID-19 looked into three variables, namely 1) number of unique employees working in skilled nursing homes, 2) number of weekly confirmed cases among residents and staff, and 3) number of weekly COVID-19 deaths among residents.

We call the variable  number of unique employees  the  independent variable , and the other two variables ( number of weekly confirmed cases among residents and staff  and  number of weekly COVID-19 deaths among residents ) as the  dependent variables .

This correlation study determined if the number of staff members in nursing homes influences the number of COVID-19 cases and deaths. It aims to understand if staffing has got to do with the transmission of the deadly coronavirus. Thus, the study’s outcome could inform policy on staffing in nursing homes during the pandemic.

A simple Pearson test may be used to correlate one variable with another variable. But the study used multiple variables. Hence, they produced  regression models  that show how multiple variables affect the outcome. Some of the variables in the study may be redundant, meaning, those variables may represent the same attribute of a population.  Stepwise multiple regression models  take care of those redundancies. Using this statistical test requires further study and experience.

Source: McGarry et al. (2021)

Topic 4: Surrounding Greenness, Stress, and Memory

Scientific evidence has shown that surrounding greenness has multiple health-related benefits. Health benefits include better cognitive functioning or better intellectual activity such as thinking, reasoning, or remembering things. These findings, however, are not well understood. A study, therefore, analyzed the relationship between surrounding greenness and memory performance, with stress as a mediating variable.

Statistical Research Question No. 4

Is there a significant relationship between exposure to and use of natural environments, stress, and memory performance?

As this article is behind a paywall and we cannot see the full article, we can content ourselves with the knowledge that three major variables were explored in this study. These are 1) exposure to and use of natural environments, 2) stress, and 3) memory performance.

Referring to the abstract of this study,  exposure to and use of natural environments  as a variable of the study may be measured in terms of the days spent by the respondent in green surroundings. That will be a ratio variable as we can count it and has an absolute zero point. Stress levels can be measured using standardized instruments like the  Perceived Stress Scale . The third variable, i.e., memory performance in terms of short-term, working memory, and overall memory may be measured using a variety of  memory assessment tools as described by Murray (2016) .

As you become more familiar and well-versed in identifying the variables you would like to investigate in your study, reading studies like this requires reading the method or methodology section. This section will tell you how the researchers measured the variables of their study. Knowing how those variables are quantified can help you design your research and formulate the appropriate statistical research questions.

Source: Lega et al. (2021)

Topic 5: Income and Happiness

This recent finding is an interesting read and is available online. Just click on the link I provide as the source below. The study sought to determine if income plays a role in people’s happiness across three age groups: young (18-30 years), middle (31-64 years), and old (65 or older). The literature review suggests that income has a positive effect on an individual’s sense of happiness. That’s because more money increases opportunities to fulfill dreams and buy more goods and services.

Reading the abstract, we can readily identify one of the variables used in the study, i.e., money. It’s easy to count that. But for happiness, that is a largely subjective matter. Happiness varies between individuals. So how did the researcher measured happiness? As previously mentioned, we need to see the methodology portion to find out why.

If you click on the link to the full text of the paper on pages 10 and 11, you will read that the researcher measured happiness using a 10-point scale. The scale was categorized into three namely, 1) unhappy, 2) happy, and 3) very happy.

An investigation was conducted to determine if the size of nursing home staff and the number of COVID-19 cases are correlated. Specifically, they looked into the number of unique employees working daily, and the outcomes include weekly counts of confirmed COVID-19 cases among residents and staff and weekly COVID-19 deaths among residents.

Statistical Research Question No. 5

Is there a significant relationship between income and happiness?

Source: Måseide (2021)

Now the statistical test used by the researcher is, honestly, beyond me. I may be able to understand it how to use it but doing so requires further study. Although I have initially did some readings on logit models, ordered logit model and generalized ordered logit model are way beyond my self-study in statistics.

Anyhow, those variables found with asterisk (***, **, and **) on page 24 tell us that there are significant relationships between income and happiness. You just have to look at the probability values and refer to the bottom of the table for the level of significance of those relationships.

I do hope that upon reaching this part of the article, you are now well familiar on how to write statistical research questions. Practice makes perfect.

References:

Chomitz, V. R., Slining, M. M., McGowan, R. J., Mitchell, S. E., Dawson, G. F., & Hacker, K. A. (2009). Is there a relationship between physical fitness and academic achievement? Positive results from public school children in the northeastern United States.  Journal of School Health ,  79 (1), 30-37.

Lega, C., Gidlow, C., Jones, M., Ellis, N., & Hurst, G. (2021). The relationship between surrounding greenness, stress and memory.  Urban Forestry & Urban Greening ,  59 , 126974.

Måseide, H. (2021). Income and Happiness: Does the relationship vary with age?

McGarry, B. E., Gandhi, A. D., Grabowski, D. C., & Barnett, M. L. (2021). Larger Nursing Home Staff Size Linked To Higher Number Of COVID-19 Cases In 2020: Study examines the relationship between staff size and COVID-19 cases in nursing homes and skilled nursing facilities. Health Affairs, 40(8), 1261-1269.

Zapata, O. (2021). The relationship between climate conditions and consumption of bottled water: A potential link between climate change and plastic pollution. Ecological Economics, 187, 107090.

© P. A. Regoniel 12 October 2021 | Updated 08 January 2024

Related Posts

The Importance of Scoping in Research

The Importance of Scoping in Research

Choosing the Right Topic: How to Find Inspiration for Your Research Paper

Choosing the Right Topic: How to Find Inspiration for Your Research Paper

How to write survey questions, about the author, patrick regoniel.

Dr. Regoniel, a faculty member of the graduate school, served as consultant to various environmental research and development projects covering issues and concerns on climate change, coral reef resources and management, economic valuation of environmental and natural resources, mining, and waste management and pollution. He has extensive experience on applied statistics, systems modelling and analysis, an avid practitioner of LaTeX, and a multidisciplinary web developer. He leverages pioneering AI-powered content creation tools to produce unique and comprehensive articles in this website.

This paper is in the following e-collection/theme issue:

Published on 17.4.2024 in Vol 26 (2024)

This is a member publication of University College London (Jisc)

Twitter Analysis of Health Care Workers’ Sentiment and Discourse Regarding Post–COVID-19 Condition in Children and Young People: Mixed Methods Study

Authors of this article:

Author Orcid Image

There are no citations yet available for this article according to Crossref .

Numbers, Facts and Trends Shaping Your World

Read our research on:

Full Topic List

Regions & Countries

  • Publications
  • Our Methods
  • Short Reads
  • Tools & Resources

Read Our Research On:

8 facts about Americans with disabilities

A bus passenger in Miami Beach, Florida, uses a handicapped access ramp. (Jeffrey Greenberg/Education Images/Universal Images Group via Getty Images)

July is Disability Pride Month in the United States, commemorating the passage of the Americans with Disabilities Act more than 30 years ago. Overall, there are about 42.5 million Americans with disabilities, making up 13% of the civilian noninstitutionalized population, according to U.S. Census Bureau data from 2021. This group includes people with hearing, vision, cognitive, walking, self-care or independent living difficulties.

Here are eight facts about Americans with disabilities, based on government data and recent Pew Research Center surveys.

Pew Research Center conducted this analysis to share key facts about Americans with disabilities for Disability Pride Month.

The analysis includes data from the U.S. Census Bureau’s American Community Survey, which  defines disability status  by asking about six types of disabilities: serious difficulty with hearing, vision, cognition, walking or climbing stairs, and difficulty with self-care and independent living. Other surveys with different definitions have estimated that a considerably larger share of Americans have disabilities .

Occupational data by disability status comes from the Bureau of Labor Statistics .

Federal education data comes from the National Center for Education Statistics. For the purposes of this analysis, disabled students include those ages 3 to 21 who are served under the federal  Individuals with Disabilities Education Act (IDEA). Through IDEA, children with disabilities are granted a free appropriate public school education and are ensured special education and related services.

Hispanic Americans in this analysis are of any race. All other racial categories include those who are not Hispanic and identify as only one race.

The public opinion findings in this analysis are based on Pew Research Center surveys. Details about each survey’s questions and methodology are available through the links in this analysis.

Due to the nature of the live telephone surveys, some Americans with disabilities are likely underrepresented in this analysis. The figures reported on technology adoption and internet use are from a phone survey that was conducted via landlines and cellphones and likely under-counted adults who are deaf or have difficulty speaking. Our surveys also do not cover those living in institutionalized group quarters, which may include some individuals who are severely disabled.

Older Americans are significantly more likely than younger adults to have a disability. Some 46% of Americans ages 75 and older and 24% of those ages 65 to 74 report having a disability, according to estimates from the Census Bureau’s 2021 American Community Survey (ACS). This compares with 12% of adults ages 35 to 64 and 8% of adults under 35.

A bar chart showing that 46% of Americans ages 75 and older report having a disability.

Americans in certain racial and ethnic groups are more likely to have a disability. American Indians and Alaska Natives (18%) are more likely than Americans of other racial and ethnic backgrounds to report having a disability, according to the 2021 ACS estimates. Asian and Hispanic Americans are least likely to say they have a disability (8% and 10%, respectively). The shares of White and Black Americans who report living with a disability fall in the middle (14% each).

The most common types of disability in the U.S. involve difficulties with walking, independent living or cognition. Some 7% of Americans report having serious ambulatory difficulties – struggling with walking or climbing stairs – according to the ACS estimates. Adults ages 75 and older and those ages 65 to 74 are the most likely to report having this kind of disability (30% and 15%, respectively). Much smaller shares of those ages 35 to 64 (6%) and those ages 18 to 34 (1%) say they have an ambulatory disability.

About 6% of Americans report difficulties with independent living – struggling to do errands alone because of physical, mental or emotional problems. And a similar share (5%) report cognitive difficulties – that is, having trouble remembering, concentrating or making decisions. Each of these disabilities is more common among older Americans than among younger age groups.

Americans with disabilities tend to earn less than those who do not have a disability. Those with a disability earned a median of $28,438 in 2021, compared with $40,948 among those without a disability, according to the Census Bureau . (These figures represent employed civilian noninstitutionalized Americans ages 16 and older. They reflect earnings in the previous 12 months in 2021 inflation-adjusted dollars.)

On average, people with disabilities accounted for 4% of employed Americans in 2022, according to the Bureau of Labor Statistics (BLS). They were most likely to be employed in management occupations (12%) and office and administrative support occupations (11%), according to annual averages compiled by the BLS, which tracks 22 occupational categories. Meanwhile, an average of about 10% of workers in transportation and material moving jobs had a disability in 2022.

Disabled Americans have lower rates of technology adoption for some devices. U.S. adults with a disability are less likely than those without a disability to say they own a desktop or laptop computer (62% vs. 81%) or a smartphone (72% vs. 88%), according to a Center survey from winter 2021 . The survey asked respondents if any “disability, handicap, or chronic disease keeps you from participating fully in work, school, housework, or other activities, or not.” (It’s important to note that there are a range of ways to measure disability in public opinion surveys.)

A dot plot showing that Americans with a disability are less likely than those without one to have traditional computer, smartphone.

Similar shares of Americans with and without disabilities say they have high-speed home internet. Even so, disabled Americans are less likely than those without a disability to report using the internet daily (75% vs. 87%). And Americans with disabilities are three times as likely as those without a disability to say they never go online (15% vs. 5%).

The percentage of U.S. public school students who receive special education or related services has increased over the last decade, according to data from the National Center for Education Statistics. During the 2021-22 school year, there were 7.3 million students receiving special education or related services in U.S. public schools , making up 15% of total enrollment. This figure rose since 2010-11, when 6.4 million disabled students made up 13% of public school enrollment.

In 2021-22, the share of disabled students in public schools varied by state, from about 20% in New York, Pennsylvania and Maine to about 12% in Idaho and Texas. These disparities are likely the result of inconsistencies in how states determine which students are eligible for special education services and some of the challenges involved with diagnosing disabilities in children.

A map that shows New York, Pennsylvania and Maine public schools serve the highest shares of students with disabilities.

Disabled Americans are much more likely than other Americans to have faced psychological distress during the COVID-19 pandemic, according to a winter 2022 Center analysis that examined survey responses from the same Americans over time.

A bar chart showing that Americans with disabilities experienced high levels of distress during the COVID-19 pandemic.

About two-thirds (66%) of adults who have a disability or health condition that keeps them from participating fully in work, school, housework or other activities reported a high level of distress at least once across four surveys conducted between March 2020 and September 2022. That compares with 34% of those who do not have a disability.

Employed Americans generally think their workplace is accessible for people with physical disabilities. Among those who don’t work fully remotely, 76% say their workplace is at least somewhat accessible for people with physical disabilities, according to a Center survey from February . This includes 51% who say it is extremely or very accessible. Another 17% say their workplace is not too or not at all accessible, while 8% are not sure.

Whether or not they consider their own workplace accessible, half of workers say they highly value physical accessibility in the workplace. Workers with disabilities are about as likely as those without disabilities to say this. (Workers are defined as those who are not self-employed and work at a company or organization with more than 10 people.)

Note: This is an update of a post originally published July 27, 2017.

  • Health Care
  • Medicine & Health

9 facts about Americans and marijuana

5 facts about black americans and health care , 5 facts about hispanic americans and health care, inflation, health costs, partisan cooperation among the nation’s top problems, by more than two-to-one, americans say medication abortion should be legal in their state, most popular.

1615 L St. NW, Suite 800 Washington, DC 20036 USA (+1) 202-419-4300 | Main (+1) 202-857-8562 | Fax (+1) 202-419-4372 |  Media Inquiries

Research Topics

  • Age & Generations
  • Coronavirus (COVID-19)
  • Economy & Work
  • Family & Relationships
  • Gender & LGBTQ
  • Immigration & Migration
  • International Affairs
  • Internet & Technology
  • Methodological Research
  • News Habits & Media
  • Non-U.S. Governments
  • Other Topics
  • Politics & Policy
  • Race & Ethnicity
  • Email Newsletters

ABOUT PEW RESEARCH CENTER  Pew Research Center is a nonpartisan fact tank that informs the public about the issues, attitudes and trends shaping the world. It conducts public opinion polling, demographic research, media content analysis and other empirical social science research. Pew Research Center does not take policy positions. It is a subsidiary of  The Pew Charitable Trusts .

Copyright 2024 Pew Research Center

Terms & Conditions

Privacy Policy

Cookie Settings

Reprints, Permissions & Use Policy

IMAGES

  1. 😊 How to write a statistical analysis paper. How to Write a Results

    example of statistical analysis research paper

  2. 6+ Research Analysis Templates

    example of statistical analysis research paper

  3. 😊 Statistical analysis paper. Free statistics project Essays and Papers

    example of statistical analysis research paper

  4. Data Analysis In Research Sample

    example of statistical analysis research paper

  5. Writing a Statistical Report for STAT 411/511 and 412/512

    example of statistical analysis research paper

  6. FREE 10+ Descriptive Research Templates in PDF

    example of statistical analysis research paper

VIDEO

  1. 2019 Mdu MCom 1st Sem Statistical Analysis for Business Question Paper

  2. How to Assess the Quantitative Data Collected from Questionnaire

  3. Demographic Analysis in SPSS

  4. Top 5 Statistical Packages for Academic Research and Analysis

  5. Research| Basic concept of research

  6. Statistics for Management I Sampling I Problems and Solutions on t test I Part 6 I Khans Commerce

COMMENTS

  1. The Beginner's Guide to Statistical Analysis

    Table of contents. Step 1: Write your hypotheses and plan your research design. Step 2: Collect data from a sample. Step 3: Summarize your data with descriptive statistics. Step 4: Test hypotheses or make estimates with inferential statistics.

  2. (PDF) An Overview of Statistical Data Analysis

    1 Introduction. Statistics is a set of methods used to analyze data. The statistic is present in all areas of science involving the. collection, handling and sorting of data, given the insight of ...

  3. Introduction to Research Statistical Analysis: An Overview of the

    Introduction. Statistical analysis is necessary for any research project seeking to make quantitative conclusions. The following is a primer for research-based statistical analysis. It is intended to be a high-level overview of appropriate statistical testing, while not diving too deep into any specific methodology.

  4. Basic statistical tools in research and data analysis

    Statistical methods involved in carrying out a study include planning, designing, collecting data, analysing, drawing meaningful interpretation and reporting of the research findings. The statistical analysis gives meaning to the meaningless numbers, thereby breathing life into a lifeless data. The results and inferences are precise only if ...

  5. PDF Anatomy of a Statistics Paper (with examples)

    12. Summary. As you read papers also notice the construction of the papers (learn from the good and bad examples). Abstract and Introduction { keys for getting readers engaged. Be gentle with your audience. Tell them your story. Writing is work { but ultimately rewarding! 13. Created Date.

  6. A Really Simple Guide to Quantitative Data Analysis

    nominal. It is important to know w hat kind of data you are planning to collect or analyse as this w ill. affect your analysis method. A 12 step approach to quantitative data analysis. Step 1 ...

  7. A practical guide to data analysis in general literature reviews

    This article is a practical guide to conducting data analysis in general literature reviews. The general literature review is a synthesis and analysis of published research on a relevant clinical issue, and is a common format for academic theses at the bachelor's and master's levels in nursing, physiotherapy, occupational therapy, public health and other related fields.

  8. Inferential Statistics

    Hypothesis testing is a formal process of statistical analysis using inferential statistics. The goal of hypothesis testing is to compare populations or assess relationships between variables using samples. Hypotheses, or predictions, are tested using statistical tests. Statistical tests also estimate sampling errors so that valid inferences ...

  9. Home

    Overview. Statistical Papers is a forum for presentation and critical assessment of statistical methods encouraging the discussion of methodological foundations and potential applications. The Journal stresses statistical methods that have broad applications, giving special attention to those relevant to the economic and social sciences.

  10. How to write statistical analysis section in medical research

    Results. Although biostatistical inputs are critical for the entire research study (online supplemental table 2), biostatistical consultations were mostly used for statistical analyses only 15.Even though the conduct of statistical analysis mismatched with the study objective and DGP was identified as the major problem in articles submitted to high-impact medical journals. 16 In addition ...

  11. PDF Structure of a Data Analysis Report

    The data analysis report isn't quite like a research paper or term paper in a class, nor like aresearch article in a journal. It is meant, primarily, to start an organized conversation between you and your client/collaborator. In that sense it is a kind of "internal" communication, sort o f like an extended memo. On the other hand it

  12. (PDF) Basic statistical tools in research and data analysis

    Abstract. Statistical methods involved in carrying out a study include planning, designing, collecting data, analysing, drawing meaningful interpretation and reporting of the research findings ...

  13. What is Statistical Analysis? Types, Methods and Examples

    Types, Methods and Examples. Statistical analysis is the process of collecting and analyzing data in order to discern patterns and trends. It is a method for removing bias from evaluating data by employing numerical analysis. This technique is useful for collecting the interpretations of research, developing statistical models, and planning ...

  14. The Beginner's Guide to Statistical Analysis

    Measuring variables. When planning a research design, you should operationalise your variables and decide exactly how you will measure them.. For statistical analysis, it's important to consider the level of measurement of your variables, which tells you what kind of data they contain:. Categorical data represents groupings. These may be nominal (e.g., gender) or ordinal (e.g. level of ...

  15. How to Write Statistics Research Paper

    Prepare and add supporting materials that will help you illustrate findings: graphs, diagrams, charts, tables, etc. You can use them in a paper's body or add them as an appendix; these elements will support your points and serve as extra proof for your findings. Last but not least: Write a concluding paragraph for your statistics research paper.

  16. Selection of Appropriate Statistical Methods for Data Analysis

    Type and distribution of the data used. For the same objective, selection of the statistical test is varying as per data types. For the nominal, ordinal, discrete data, we use nonparametric methods while for continuous data, parametric methods as well as nonparametric methods are used.[] For example, in the regression analysis, when our outcome variable is categorical, logistic regression ...

  17. How To Write a Statistical Research Paper: Tips, Topics, Outline

    Working on a statistics paper can be quite challenging to work on. But with the right information sources, everything becomes easy. This guide will help you reveal the secret of preparing such essays. Also, don't forget to do more reading to broaden your knowledge. You can find statistics research paper examples and refer to them for ideas.

  18. Statistics Research Paper

    View sample Statistics Research Paper. Browse other research paper examples and check the list of research paper topics for more inspiration. If you need a re ... It is widely recognized that any statistical analysis can only be as good as the underlying data. Consequently, statisticians take great care in the the design of methods for data ...

  19. Statistical Research Questions: Five Examples for Quantitative Analysis

    Introduction. Five Examples of Statistical Research Questions. Topic 1: Physical Fitness and Academic Achievement. Statistical Research Question No. 1. Topic 2: Climate Conditions and Consumption of Bottled Water. Statistical Research Question No. 2. Topic 3: Nursing Home Staff Size and Number of COVID-19 Cases.

  20. Journal of Medical Internet Research

    Background: The COVID-19 pandemic has had a significant global impact, with millions of cases and deaths. Research highlights the persistence of symptoms over time (post-COVID-19 condition), a situation of particular concern in children and young people with symptoms. Social media such as Twitter (subsequently rebranded as X) could provide valuable information on the impact of the post ...

  21. Reporting Statistics in APA Style

    To report the results of a correlation, include the following: the degrees of freedom in parentheses. the r value (the correlation coefficient) the p value. Example: Reporting correlation results. We found a strong correlation between average temperature and new daily cases of COVID-19, r (357) = .42, p < .001.

  22. What the data says about abortion in the U.S.

    The CDC says that in 2021, there were 11.6 abortions in the U.S. per 1,000 women ages 15 to 44. (That figure excludes data from California, the District of Columbia, Maryland, New Hampshire and New Jersey.) Like Guttmacher's data, the CDC's figures also suggest a general decline in the abortion rate over time.

  23. 8 facts about Americans with disabilities

    Pew Research Center conducted this analysis to share key facts about Americans with disabilities for Disability Pride Month. The analysis includes data from the U.S. Census Bureau's American Community Survey, which defines disability status by asking about six types of disabilities: serious difficulty with hearing, vision, cognition, walking or climbing stairs, and difficulty with self-care ...