2. variables
3. variables
4. variables
5. variables
6. variables
7. variables
8. variables
The simplest way to understand a variable is as any characteristic or attribute that can experience change or vary over time or context – hence the name “variable”. For example, the dosage of a particular medicine could be classified as a variable, as the amount can vary (i.e., a higher dose or a lower dose). Similarly, gender, age or ethnicity could be considered demographic variables, because each person varies in these respects.
Within research, especially scientific research, variables form the foundation of studies, as researchers are often interested in how one variable impacts another, and the relationships between different variables. For example:
As you can see, variables are often used to explain relationships between different elements and phenomena. In scientific studies, especially experimental studies, the objective is often to understand the causal relationships between variables. In other words, the role of cause and effect between variables. This is achieved by manipulating certain variables while controlling others – and then observing the outcome. But, we’ll get into that a little later…
Variables can be a little intimidating for new researchers because there are a wide variety of variables, and oftentimes, there are multiple labels for the same thing. To lay a firm foundation, we’ll first look at the three main types of variables, namely:
Simply put, the independent variable is the “ cause ” in the relationship between two (or more) variables. In other words, when the independent variable changes, it has an impact on another variable.
For example:
It’s useful to know that independent variables can go by a few different names, including, explanatory variables (because they explain an event or outcome) and predictor variables (because they predict the value of another variable). Terminology aside though, the most important takeaway is that independent variables are assumed to be the “cause” in any cause-effect relationship. As you can imagine, these types of variables are of major interest to researchers, as many studies seek to understand the causal factors behind a phenomenon.
While the independent variable is the “ cause ”, the dependent variable is the “ effect ” – or rather, the affected variable . In other words, the dependent variable is the variable that is assumed to change as a result of a change in the independent variable.
Keeping with the previous example, let’s look at some dependent variables in action:
In scientific studies, researchers will typically pay very close attention to the dependent variable (or variables), carefully measuring any changes in response to hypothesised independent variables. This can be tricky in practice, as it’s not always easy to reliably measure specific phenomena or outcomes – or to be certain that the actual cause of the change is in fact the independent variable.
As the adage goes, correlation is not causation . In other words, just because two variables have a relationship doesn’t mean that it’s a causal relationship – they may just happen to vary together. For example, you could find a correlation between the number of people who own a certain brand of car and the number of people who have a certain type of job. Just because the number of people who own that brand of car and the number of people who have that type of job is correlated, it doesn’t mean that owning that brand of car causes someone to have that type of job or vice versa. The correlation could, for example, be caused by another factor such as income level or age group, which would affect both car ownership and job type.
To confidently establish a causal relationship between an independent variable and a dependent variable (i.e., X causes Y), you’ll typically need an experimental design , where you have complete control over the environmen t and the variables of interest. But even so, this doesn’t always translate into the “real world”. Simply put, what happens in the lab sometimes stays in the lab!
As an alternative to pure experimental research, correlational or “ quasi-experimental ” research (where the researcher cannot manipulate or change variables) can be done on a much larger scale more easily, allowing one to understand specific relationships in the real world. These types of studies also assume some causality between independent and dependent variables, but it’s not always clear. So, if you go this route, you need to be cautious in terms of how you describe the impact and causality between variables and be sure to acknowledge any limitations in your own research.
In an experimental design, a control variable (or controlled variable) is a variable that is intentionally held constant to ensure it doesn’t have an influence on any other variables. As a result, this variable remains unchanged throughout the course of the study. In other words, it’s a variable that’s not allowed to vary – tough life 🙂
As we mentioned earlier, one of the major challenges in identifying and measuring causal relationships is that it’s difficult to isolate the impact of variables other than the independent variable. Simply put, there’s always a risk that there are factors beyond the ones you’re specifically looking at that might be impacting the results of your study. So, to minimise the risk of this, researchers will attempt (as best possible) to hold other variables constant . These factors are then considered control variables.
Some examples of variables that you may need to control include:
Which specific variables need to be controlled for will vary tremendously depending on the research project at hand, so there’s no generic list of control variables to consult. As a researcher, you’ll need to think carefully about all the factors that could vary within your research context and then consider how you’ll go about controlling them. A good starting point is to look at previous studies similar to yours and pay close attention to which variables they controlled for.
Of course, you won’t always be able to control every possible variable, and so, in many cases, you’ll just have to acknowledge their potential impact and account for them in the conclusions you draw. Every study has its limitations , so don’t get fixated or discouraged by troublesome variables. Nevertheless, always think carefully about the factors beyond what you’re focusing on – don’t make assumptions!
As we mentioned, independent, dependent and control variables are the most common variables you’ll come across in your research, but they’re certainly not the only ones you need to be aware of. Next, we’ll look at a few “secondary” variables that you need to keep in mind as you design your research.
Let’s jump into it…
A moderating variable is a variable that influences the strength or direction of the relationship between an independent variable and a dependent variable. In other words, moderating variables affect how much (or how little) the IV affects the DV, or whether the IV has a positive or negative relationship with the DV (i.e., moves in the same or opposite direction).
For example, in a study about the effects of sleep deprivation on academic performance, gender could be used as a moderating variable to see if there are any differences in how men and women respond to a lack of sleep. In such a case, one may find that gender has an influence on how much students’ scores suffer when they’re deprived of sleep.
It’s important to note that while moderators can have an influence on outcomes , they don’t necessarily cause them ; rather they modify or “moderate” existing relationships between other variables. This means that it’s possible for two different groups with similar characteristics, but different levels of moderation, to experience very different results from the same experiment or study design.
Mediating variables are often used to explain the relationship between the independent and dependent variable (s). For example, if you were researching the effects of age on job satisfaction, then education level could be considered a mediating variable, as it may explain why older people have higher job satisfaction than younger people – they may have more experience or better qualifications, which lead to greater job satisfaction.
Mediating variables also help researchers understand how different factors interact with each other to influence outcomes. For instance, if you wanted to study the effect of stress on academic performance, then coping strategies might act as a mediating factor by influencing both stress levels and academic performance simultaneously. For example, students who use effective coping strategies might be less stressed but also perform better academically due to their improved mental state.
In addition, mediating variables can provide insight into causal relationships between two variables by helping researchers determine whether changes in one factor directly cause changes in another – or whether there is an indirect relationship between them mediated by some third factor(s). For instance, if you wanted to investigate the impact of parental involvement on student achievement, you would need to consider family dynamics as a potential mediator, since it could influence both parental involvement and student achievement simultaneously.
A confounding variable (also known as a third variable or lurking variable ) is an extraneous factor that can influence the relationship between two variables being studied. Specifically, for a variable to be considered a confounding variable, it needs to meet two criteria:
Some common examples of confounding variables include demographic factors such as gender, ethnicity, socioeconomic status, age, education level, and health status. In addition to these, there are also environmental factors to consider. For example, air pollution could confound the impact of the variables of interest in a study investigating health outcomes.
Naturally, it’s important to identify as many confounding variables as possible when conducting your research, as they can heavily distort the results and lead you to draw incorrect conclusions . So, always think carefully about what factors may have a confounding effect on your variables of interest and try to manage these as best you can.
Latent variables are unobservable factors that can influence the behaviour of individuals and explain certain outcomes within a study. They’re also known as hidden or underlying variables , and what makes them rather tricky is that they can’t be directly observed or measured . Instead, latent variables must be inferred from other observable data points such as responses to surveys or experiments.
For example, in a study of mental health, the variable “resilience” could be considered a latent variable. It can’t be directly measured , but it can be inferred from measures of mental health symptoms, stress, and coping mechanisms. The same applies to a lot of concepts we encounter every day – for example:
One way in which we overcome the challenge of measuring the immeasurable is latent variable models (LVMs). An LVM is a type of statistical model that describes a relationship between observed variables and one or more unobserved (latent) variables. These models allow researchers to uncover patterns in their data which may not have been visible before, thanks to their complexity and interrelatedness with other variables. Those patterns can then inform hypotheses about cause-and-effect relationships among those same variables which were previously unknown prior to running the LVM. Powerful stuff, we say!
In the world of scientific research, there’s no shortage of variable types, some of which have multiple names and some of which overlap with each other. In this post, we’ve covered some of the popular ones, but remember that this is not an exhaustive list .
To recap, we’ve explored:
If you’re still feeling a bit lost and need a helping hand with your research project, check out our 1-on-1 coaching service , where we guide you through each step of the research journey. Also, be sure to check out our free dissertation writing course and our collection of free, fully-editable chapter templates .
This post was based on one of our popular Research Bootcamps . If you're working on a research project, you'll definitely want to check this out ...
Very informative, concise and helpful. Thank you
Helping information.Thanks
practical and well-demonstrated
Very helpful and insightful
Your email address will not be published. Required fields are marked *
Save my name, email, and website in this browser for the next time I comment.
Saul McLeod, PhD
Editor-in-Chief for Simply Psychology
BSc (Hons) Psychology, MRes, PhD, University of Manchester
Saul McLeod, PhD., is a qualified psychology teacher with over 18 years of experience in further and higher education. He has been published in peer-reviewed journals, including the Journal of Clinical Psychology.
Learn about our Editorial Process
Olivia Guy-Evans, MSc
Associate Editor for Simply Psychology
BSc (Hons) Psychology, MSc Psychology of Education
Olivia Guy-Evans is a writer and associate editor for Simply Psychology. She has previously worked in healthcare and educational sectors.
On This Page:
A research hypothesis, in its plural form “hypotheses,” is a specific, testable prediction about the anticipated results of a study, established at its outset. It is a key component of the scientific method .
Hypotheses connect theory to data and guide the research process towards expanding scientific understanding
Predictions typically arise from a thorough knowledge of the research literature, curiosity about real-world problems or implications, and integrating this to advance theory. They build on existing literature while providing new insight.
Alternative hypothesis.
The research hypothesis is often called the alternative or experimental hypothesis in experimental research.
It typically suggests a potential relationship between two key variables: the independent variable, which the researcher manipulates, and the dependent variable, which is measured based on those changes.
The alternative hypothesis states a relationship exists between the two variables being studied (one variable affects the other).
A hypothesis is a testable statement or prediction about the relationship between two or more variables. It is a key component of the scientific method. Some key points about hypotheses:
In summary, a hypothesis is a precise, testable statement of what researchers expect to happen in a study and why. Hypotheses connect theory to data and guide the research process towards expanding scientific understanding.
An experimental hypothesis predicts what change(s) will occur in the dependent variable when the independent variable is manipulated.
It states that the results are not due to chance and are significant in supporting the theory being investigated.
The alternative hypothesis can be directional, indicating a specific direction of the effect, or non-directional, suggesting a difference without specifying its nature. It’s what researchers aim to support or demonstrate through their study.
The null hypothesis states no relationship exists between the two variables being studied (one variable does not affect the other). There will be no changes in the dependent variable due to manipulating the independent variable.
It states results are due to chance and are not significant in supporting the idea being investigated.
The null hypothesis, positing no effect or relationship, is a foundational contrast to the research hypothesis in scientific inquiry. It establishes a baseline for statistical testing, promoting objectivity by initiating research from a neutral stance.
Many statistical methods are tailored to test the null hypothesis, determining the likelihood of observed results if no true effect exists.
This dual-hypothesis approach provides clarity, ensuring that research intentions are explicit, and fosters consistency across scientific studies, enhancing the standardization and interpretability of research outcomes.
A non-directional hypothesis, also known as a two-tailed hypothesis, predicts that there is a difference or relationship between two variables but does not specify the direction of this relationship.
It merely indicates that a change or effect will occur without predicting which group will have higher or lower values.
For example, “There is a difference in performance between Group A and Group B” is a non-directional hypothesis.
A directional (one-tailed) hypothesis predicts the nature of the effect of the independent variable on the dependent variable. It predicts in which direction the change will take place. (i.e., greater, smaller, less, more)
It specifies whether one variable is greater, lesser, or different from another, rather than just indicating that there’s a difference without specifying its nature.
For example, “Exercise increases weight loss” is a directional hypothesis.
The Falsification Principle, proposed by Karl Popper , is a way of demarcating science from non-science. It suggests that for a theory or hypothesis to be considered scientific, it must be testable and irrefutable.
Falsifiability emphasizes that scientific claims shouldn’t just be confirmable but should also have the potential to be proven wrong.
It means that there should exist some potential evidence or experiment that could prove the proposition false.
However many confirming instances exist for a theory, it only takes one counter observation to falsify it. For example, the hypothesis that “all swans are white,” can be falsified by observing a black swan.
For Popper, science should attempt to disprove a theory rather than attempt to continually provide evidence to support a research hypothesis.
Hypotheses make probabilistic predictions. They state the expected outcome if a particular relationship exists. However, a study result supporting a hypothesis does not definitively prove it is true.
All studies have limitations. There may be unknown confounding factors or issues that limit the certainty of conclusions. Additional studies may yield different results.
In science, hypotheses can realistically only be supported with some degree of confidence, not proven. The process of science is to incrementally accumulate evidence for and against hypothesized relationships in an ongoing pursuit of better models and explanations that best fit the empirical data. But hypotheses remain open to revision and rejection if that is where the evidence leads.
We can never 100% prove the alternative hypothesis. Instead, we see if we can disprove, or reject the null hypothesis.
If we reject the null hypothesis, this doesn’t mean that our alternative hypothesis is correct but does support the alternative/experimental hypothesis.
Upon analysis of the results, an alternative hypothesis can be rejected or supported, but it can never be proven to be correct. We must avoid any reference to results proving a theory as this implies 100% certainty, and there is always a chance that evidence may exist which could refute a theory.
Consider a hypothesis many teachers might subscribe to: students work better on Monday morning than on Friday afternoon (IV=Day, DV= Standard of work).
Now, if we decide to study this by giving the same group of students a lesson on a Monday morning and a Friday afternoon and then measuring their immediate recall of the material covered in each session, we would end up with the following:
Run a free plagiarism check in 10 minutes, automatically generate references for free.
Published on 6 May 2022 by Shona McCombes .
A hypothesis is a statement that can be tested by scientific research. If you want to test a relationship between two or more variables, you need to write hypotheses before you start your experiment or data collection.
What is a hypothesis, developing a hypothesis (with example), hypothesis examples, frequently asked questions about writing hypotheses.
A hypothesis states your predictions about what your research will find. It is a tentative answer to your research question that has not yet been tested. For some research projects, you might have to write several hypotheses that address different aspects of your research question.
A hypothesis is not just a guess – it should be based on existing theories and knowledge. It also has to be testable, which means you can support or refute it through scientific research methods (such as experiments, observations, and statistical analysis of data).
Hypotheses propose a relationship between two or more variables . An independent variable is something the researcher changes or controls. A dependent variable is something the researcher observes and measures.
In this example, the independent variable is exposure to the sun – the assumed cause . The dependent variable is the level of happiness – the assumed effect .
Step 1: ask a question.
Writing a hypothesis begins with a research question that you want to answer. The question should be focused, specific, and researchable within the constraints of your project.
Your initial answer to the question should be based on what is already known about the topic. Look for theories and previous studies to help you form educated assumptions about what your research will find.
At this stage, you might construct a conceptual framework to identify which variables you will study and what you think the relationships are between them. Sometimes, you’ll have to operationalise more complex constructs.
Now you should have some idea of what you expect to find. Write your initial answer to the question in a clear, concise sentence.
You need to make sure your hypothesis is specific and testable. There are various ways of phrasing a hypothesis, but all the terms you use should have clear definitions, and the hypothesis should contain:
To identify the variables, you can write a simple prediction in if … then form. The first part of the sentence states the independent variable and the second part states the dependent variable.
In academic research, hypotheses are more commonly phrased in terms of correlations or effects, where you directly state the predicted relationship between variables.
If you are comparing two groups, the hypothesis can state what difference you expect to find between them.
If your research involves statistical hypothesis testing , you will also have to write a null hypothesis. The null hypothesis is the default position that there is no association between the variables. The null hypothesis is written as H 0 , while the alternative hypothesis is H 1 or H a .
Research question | Hypothesis | Null hypothesis |
---|---|---|
What are the health benefits of eating an apple a day? | Increasing apple consumption in over-60s will result in decreasing frequency of doctor’s visits. | Increasing apple consumption in over-60s will have no effect on frequency of doctor’s visits. |
Which airlines have the most delays? | Low-cost airlines are more likely to have delays than premium airlines. | Low-cost and premium airlines are equally likely to have delays. |
Can flexible work arrangements improve job satisfaction? | Employees who have flexible working hours will report greater job satisfaction than employees who work fixed hours. | There is no relationship between working hour flexibility and job satisfaction. |
How effective is secondary school sex education at reducing teen pregnancies? | Teenagers who received sex education lessons throughout secondary school will have lower rates of unplanned pregnancy than teenagers who did not receive any sex education. | Secondary school sex education has no effect on teen pregnancy rates. |
What effect does daily use of social media have on the attention span of under-16s? | There is a negative correlation between time spent on social media and attention span in under-16s. | There is no relationship between social media use and attention span in under-16s. |
Hypothesis testing is a formal procedure for investigating our ideas about the world using statistics. It is used by scientists to test specific predictions, called hypotheses , by calculating how likely it is that a pattern or relationship between variables could have arisen by chance.
A hypothesis is not just a guess. It should be based on existing theories and knowledge. It also has to be testable, which means you can support or refute it through scientific research methods (such as experiments, observations, and statistical analysis of data).
A research hypothesis is your proposed answer to your research question. The research hypothesis usually includes an explanation (‘ x affects y because …’).
A statistical hypothesis, on the other hand, is a mathematical statement about a population parameter. Statistical hypotheses always come in pairs: the null and alternative hypotheses. In a well-designed study , the statistical hypotheses correspond logically to the research hypothesis.
If you want to cite this source, you can copy and paste the citation or click the ‘Cite this Scribbr article’ button to automatically add the citation to our free Reference Generator.
McCombes, S. (2022, May 06). How to Write a Strong Hypothesis | Guide & Examples. Scribbr. Retrieved 16 September 2024, from https://www.scribbr.co.uk/research-methods/hypothesis-writing/
Other students also liked, operationalisation | a guide with examples, pros & cons, what is a conceptual framework | tips & examples, a quick guide to experimental design | 5 steps & examples.
Statistics By Jim
Making statistics intuitive
By Jim Frost 15 Comments
In this post, learn the definitions of independent and dependent variables, how to identify each type, how they differ between different types of studies, and see examples of them in use.
Independent variables (IVs) are the ones that you include in the model to explain or predict changes in the dependent variable. The name helps you understand their role in statistical analysis. These variables are independent . In this context, independent indicates that they stand alone and other variables in the model do not influence them. The researchers are not seeking to understand what causes the independent variables to change.
Independent variables are also known as predictors, factors , treatment variables, explanatory variables, input variables, x-variables, and right-hand variables—because they appear on the right side of the equals sign in a regression equation. In notation, statisticians commonly denote them using Xs. On graphs, analysts place independent variables on the horizontal, or X, axis.
In machine learning, independent variables are known as features.
For example, in a plant growth study, the independent variables might be soil moisture (continuous) and type of fertilizer (categorical).
Statistical models will estimate effect sizes for the independent variables.
Relate post : Effect Sizes in Statistics
The nature of independent variables changes based on the type of experiment or study:
Controlled experiments : Researchers systematically control and set the values of the independent variables. In randomized experiments, relationships between independent and dependent variables tend to be causal. The independent variables cause changes in the dependent variable.
Observational studies : Researchers do not set the values of the explanatory variables but instead observe them in their natural environment. When the independent and dependent variables are correlated, those relationships might not be causal.
When you include one independent variable in a regression model, you are performing simple regression. For more than one independent variable, it is multiple regression. Despite the different names, it’s really the same analysis with the same interpretations and assumptions.
Determining which IVs to include in a statistical model is known as model specification. That process involves in-depth research and many subject-area, theoretical, and statistical considerations. At its most basic level, you’ll want to include the predictors you are specifically assessing in your study and confounding variables that will bias your results if you don’t add them—particularly for observational studies.
For more information about choosing independent variables, read my post about Specifying the Correct Regression Model .
Related posts : Randomized Experiments , Observational Studies , Covariates , and Confounding Variables
The dependent variable (DV) is what you want to use the model to explain or predict. The values of this variable depend on other variables. It is the outcome that you’re studying. It’s also known as the response variable, outcome variable, and left-hand variable. Statisticians commonly denote them using a Y. Traditionally, graphs place dependent variables on the vertical, or Y, axis.
For example, in the plant growth study example, a measure of plant growth is the dependent variable. That is the outcome of the experiment, and we want to determine what affects it.
If you’re reading a study’s write-up, how do you distinguish independent variables from dependent variables? Here are some tips!
How statisticians discuss independent variables changes depending on the field of study and type of experiment.
In randomized experiments, look for the following descriptions to identify the independent variables:
In observational studies, independent variables are a bit different. While the researchers likely want to establish causation, that’s harder to do with this type of study, so they often won’t use the word “cause.” They also don’t set the values of the predictors. Some independent variables are the experiment’s focus, while others help keep the experimental results valid.
Here’s how to recognize independent variables in observational studies:
Regardless of the study type, if you see an estimated effect size, it is an independent variable.
Dependent variables are the outcome. The IVs explain the variability or causes changes in the DV. Focus on the “depends” aspect. The value of the dependent variable depends on the IVs. If Y depends on X, then Y is the dependent variable. This aspect applies to both randomized experiments and observational studies.
In an observational study about the effects of smoking, the researchers observe the subjects’ smoking status (smoker/non-smoker) and their lung cancer rates. It’s an observational study because they cannot randomly assign subjects to either the smoking or non-smoking group. In this study, the researchers want to know whether lung cancer rates depend on smoking status. Therefore, the lung cancer rate is the dependent variable.
In a randomized COVID-19 vaccine experiment , the researchers randomly assign subjects to the treatment or control group. They want to determine whether COVID-19 infection rates depend on vaccination status. Hence, the infection rate is the DV.
Note that a variable can be an independent variable in one study but a dependent variable in another. It depends on the context.
For example, one study might assess how the amount of exercise (IV) affects health (DV). However, another study might study the factors (IVs) that influence how much someone exercises (DV). The amount of exercise is an independent variable in one study but a dependent variable in the other!
Regression analysis and ANOVA mathematically describe the relationships between each independent variable and the dependent variable. Typically, you want to determine how changes in one or more predictors associate with changes in the dependent variable. These analyses estimate an effect size for each independent variable.
Suppose researchers study the relationship between wattage, several types of filaments, and the output from a light bulb. In this study, light output is the dependent variable because it depends on the other two variables. Wattage (continuous) and filament type (categorical) are the independent variables.
After performing the regression analysis, the researchers will understand the nature of the relationship between these variables. How much does the light output increase on average for each additional watt? Does the mean light output differ by filament types? They will also learn whether these effects are statistically significant.
Related post : When to Use Regression Analysis
As I mentioned earlier, graphs traditionally display the independent variables on the horizontal X-axis and the dependent variable on the vertical Y-axis. The type of graph depends on the nature of the variables. Here are a couple of examples.
Suppose you experiment to determine whether various teaching methods affect learning outcomes. Teaching method is a categorical predictor that defines the experimental groups. To display this type of data, you can use a boxplot, as shown below.
The groups are along the horizontal axis, while the dependent variable, learning outcomes, is on the vertical. From the graph, method 4 has the best results. A one-way ANOVA will tell you whether these results are statistically significant. Learn more about interpreting boxplots .
Now, imagine that you are studying people’s height and weight. Specifically, do height increases cause weight to increase? Consequently, height is the independent variable on the horizontal axis, and weight is the dependent variable on the vertical axis. You can use a scatterplot to display this type of data.
It appears that as height increases, weight tends to increase. Regression analysis will tell you if these results are statistically significant. Learn more about interpreting scatterplots .
April 2, 2024 at 2:05 am
Hi again Jim
Thanks so much for taking an interest in New Zealand’s Equity Index.
Rather than me trying to explain what our Ministry of Education has done, here is a link to a fairly short paper. Scroll down to page 4 of this (if you have the inclination) – https://fyi.org.nz/request/21253/response/80708/attach/4/1301098%20Response%20and%20Appendix.pdf
The Equity Index is used to allocate only 4% of total school funding. The most advantaged 5% of schools get no “equity funding” and the other 95% get a share of the equity funding pool based on their index score. We are talking a maximum of around $1,000NZD per child per year for the most disadvantaged schools. The average amount is around $200-$300 per child per year.
My concern is that I thought the dependent variable is the thing you want to explain or predict using one or more independent variables. Choosing the form of dependent variable that gets a good fit seems to be answering the question “what can we predict well?” rather than “how do we best predict the factor of interest?” The factor is educational achievement and I think this should have been decided upon using theory rather than experimentation with the data.
As it turns out, the Ministry has chosen a measure of educational achievement that puts a heavy weight on achieving an “excellence” rating on a qualification and a much lower weight on simply gaining a qualification. My reading is that they have taken what our universities do when looking at which students to admit.
It doesn’t seem likely to me that a heavy weighting on excellent achievement is appropriate for targeting extra funding to schools with a lot of under-achieving students.
However, my stats knowledge isn’t extensive and it’s definitely rusty, so your thoughts are most helpful.
Regards Kathy Spencer
April 1, 2024 at 4:08 pm
Hi Jim, Great website, thank you.
I have been looking at New Zealand’s Equity Index which is used to allocate a small amount of extra funding to schools attended by children from disadvantaged backgrounds. The Index uses 37 socioeconomic measures relating to a child’s and their parents’ backgrounds that are found to be associated with educational achievement.
I was a bit surprised to read how they had decided on the dependent variable to be used as the measure of educational achievement, or dependent variable. Part of the process was as follows- “Each measure was tested to see the degree to which it could be predicted by the socioeconomic factors selected for the Equity Index.”
Any comment?
Many thanks Kathy Spencer
April 1, 2024 at 9:20 pm
That’s a very complex study and I don’t know much about it. So, that limits what I can say about it. But I’ll give you a few thoughts that come to mind.
This method is common in educational and social research, particularly when the goal is to understand or mitigate the impact of socioeconomic disparities on educational outcomes.
There are the usual concerns about not confusing correlation with causation. However, because this program seems to quantify barriers and then provide extra funding based on the index, I don’t think that’s a problem. They’re not attempting to adjust the socioeconomic measures so no worries about whether they’re directly causal or not.
I might have a small concern about cherry picking the model that happens to maximize the R-squared. Chasing the R-squared rather than having theory drive model selecting is often problematic. Chasing the best fit increases the likelihood that the model fits this specific dataset best by random chance rather than being truly the best. If so, it won’t perform as well outside the dataset used to fit the model. Hopefully, they validated the predicted ability of the model using other data.
However, I’m not sure if the extra funding is determined by the model? I don’t know if the index value is calculated separately outside the candidate models and then fed into the various models. Or does the choice of model affect how the index value is calculated? If it’s the former, then the funding doesn’t depend on a potentially cherry picked model. If the latter, it does.
So, I’m not really clear on the purpose of the model. I’m guessing they just want to validate their Equity Index. And maximizing the R-squared doesn’t really say it’s the best Index but it does at least show that it likely has some merit. I’d be curious how the took the 37 measures and combined them to one index. So, I have more questions than answers. I don’t mean that in a critical sense. Just that I know almost nothing about this program.
I’m curious, what was the outcome they picked? How high was the R-squared? And what were your concerns?
February 6, 2024 at 6:57 pm
Excellent explanation, thank you.
February 5, 2024 at 5:04 pm
Thank you for this insightful blog. Is it valid to use a dependent variable delivered from the mean of independent variables in multiple regression if you want to evaluate the influence of each unique independent variable on the dependent variables?
February 5, 2024 at 11:11 pm
It’s difficult to answer your question because I’m not sure what you mean that the DV is “delivered from the mean of IVs.” If you mean that multiple IVs explain changes in the DV’s mean, yes, that’s the standard use for multiple regression.
If you mean something else, please explain in further detail. Thanks!
February 6, 2024 at 6:32 am
What I meant is; the DV values used as parameters for multiple regression is basically calculated as the average of the IVs. For instance:
From 3 IVs (X1, X2, X3), Y is delivered as :
Y = (Sum of all IVs) / (3)
Then the resulting Y is used as the DV along with the initial IVs to compute the multiple regression.
February 6, 2024 at 2:17 pm
There are a couple of reasons why you shouldn’t do that.
For starters, Y-hat (the predicted value of the regression equation) is the mean of the DV given specific values of the IV. However, that mean is calculated by using the regression coefficients and constant in the regression equation. You don’t calculate the DV mean as the sum of the IVs divided by the number of IVs. Perhaps given a very specific subject-area context, using this approach might seem to make sense but there are other problems.
A critical problem is that the Y is now calculated using the IVs. Instead, the DVs should be measured outcomes and not calculated from IVs. This violates regression assumptions and produces questionable results.
Additionally, it complicates the interpretation. Because the DV is calculated from the IV, you know the regression analysis will find a relationship between them. But you have no idea if that relationship exists in the real world. This complication occurs because your results are based on forcing the DV to equal a function of the IVs and do not reflect real-world outcomes.
In short, DVs should be real-world outcomes that you measure! And be sure to keep your IVs and DV independent. Let the regression analysis estimate the regression equation from your data that contains measured DVs. Don’t use a function to force the DV to equal some function of the IVs because that’s the opposite direction of how regression works!
I hope that helps!
September 6, 2022 at 7:43 pm
Thank you for sharing.
March 3, 2022 at 1:59 am
Excellent explanation.
February 13, 2022 at 12:31 pm
Thanks a lot for creating this excellent blog. This is my go-to resource for Statistics.
I had been pondering over a question for sometime, it would be great if you could shed some light on this.
In linear and non-linear regression, should the distribution of independent and dependent variables be unskewed? When is there a need to transform the data (say, Box-Cox transformation), and do we transform the independent variables as well?
October 28, 2021 at 12:55 pm
If I use a independent variable (X) and it displays a low p-value <.05, why is it if I introduce another independent variable to regression the coefficient and p-value of Y that I used in first regression changes to look insignificant? The second variable that I introduced has a low p-value in regression.
October 29, 2021 at 11:22 pm
Keep in mind that the significance of each IV is calculated after accounting for the variance of all the other variables in the model, assuming you’re using the standard adjusted sums of squares rather than sequential sums of squares. The sums of squares (SS) is a measure of how much dependent variable variability that each IV accounts for. In the illustration below, I’ll assume you’re using the standard of adjusted SS.
So, let’s say that originally you have X1 in the model along with some other IVs. Your model estimates the significance of X1 after assessing the variability that the other IVs account for and finds that X1 is significant. Now, you add X2 to the model in addition to X1 and the other IVs. Now, when assessing X1, the model accounts for the variability of the IVs including the newly added X2. And apparently X2 explains a good portion of the variability. X1 is no longer able to account for that variability, which causes it to not be statistically significant.
In other words, X2 explains some of the variability that X1 previously explained. Because X1 no longer explains it, it is no longer significant.
Additionally, the significance of IVs is more likely to change when you add or remove IVs that are correlated. Correlated IVs is known as multicollinearity. Multicollinearity can be a problem when you have too much. Given the change in significance, I’d check your model for multicollinearity just to be safe! Click the link to read a post that wrote about that!
September 6, 2021 at 8:35 am
nice explanation
August 25, 2021 at 3:09 am
it is excellent explanation
Hypothesis Definition, Format, Examples, and Tips
Verywell / Alex Dos Diaz
Falsifiability of a hypothesis.
Hypotheses examples.
A hypothesis is a tentative statement about the relationship between two or more variables. It is a specific, testable prediction about what you expect to happen in a study. It is a preliminary answer to your question that helps guide the research process.
Consider a study designed to examine the relationship between sleep deprivation and test performance. The hypothesis might be: "This study is designed to assess the hypothesis that sleep-deprived people will perform worse on a test than individuals who are not sleep-deprived."
A hypothesis is crucial to scientific research because it offers a clear direction for what the researchers are looking to find. This allows them to design experiments to test their predictions and add to our scientific knowledge about the world. This article explores how a hypothesis is used in psychology research, how to write a good hypothesis, and the different types of hypotheses you might use.
In the scientific method , whether it involves research in psychology, biology, or some other area, a hypothesis represents what the researchers think will happen in an experiment. The scientific method involves the following steps:
The hypothesis is a prediction, but it involves more than a guess. Most of the time, the hypothesis begins with a question which is then explored through background research. At this point, researchers then begin to develop a testable hypothesis.
Unless you are creating an exploratory study, your hypothesis should always explain what you expect to happen.
In a study exploring the effects of a particular drug, the hypothesis might be that researchers expect the drug to have some type of effect on the symptoms of a specific illness. In psychology, the hypothesis might focus on how a certain aspect of the environment might influence a particular behavior.
Remember, a hypothesis does not have to be correct. While the hypothesis predicts what the researchers expect to see, the goal of the research is to determine whether this guess is right or wrong. When conducting an experiment, researchers might explore numerous factors to determine which ones might contribute to the ultimate outcome.
In many cases, researchers may find that the results of an experiment do not support the original hypothesis. When writing up these results, the researchers might suggest other options that should be explored in future studies.
In many cases, researchers might draw a hypothesis from a specific theory or build on previous research. For example, prior research has shown that stress can impact the immune system. So a researcher might hypothesize: "People with high-stress levels will be more likely to contract a common cold after being exposed to the virus than people who have low-stress levels."
In other instances, researchers might look at commonly held beliefs or folk wisdom. "Birds of a feather flock together" is one example of folk adage that a psychologist might try to investigate. The researcher might pose a specific hypothesis that "People tend to select romantic partners who are similar to them in interests and educational level."
So how do you write a good hypothesis? When trying to come up with a hypothesis for your research or experiments, ask yourself the following questions:
Before you come up with a specific hypothesis, spend some time doing background research. Once you have completed a literature review, start thinking about potential questions you still have. Pay attention to the discussion section in the journal articles you read . Many authors will suggest questions that still need to be explored.
To form a hypothesis, you should take these steps:
In the scientific method , falsifiability is an important part of any valid hypothesis. In order to test a claim scientifically, it must be possible that the claim could be proven false.
Students sometimes confuse the idea of falsifiability with the idea that it means that something is false, which is not the case. What falsifiability means is that if something was false, then it is possible to demonstrate that it is false.
One of the hallmarks of pseudoscience is that it makes claims that cannot be refuted or proven false.
A variable is a factor or element that can be changed and manipulated in ways that are observable and measurable. However, the researcher must also define how the variable will be manipulated and measured in the study.
Operational definitions are specific definitions for all relevant factors in a study. This process helps make vague or ambiguous concepts detailed and measurable.
For example, a researcher might operationally define the variable " test anxiety " as the results of a self-report measure of anxiety experienced during an exam. A "study habits" variable might be defined by the amount of studying that actually occurs as measured by time.
These precise descriptions are important because many things can be measured in various ways. Clearly defining these variables and how they are measured helps ensure that other researchers can replicate your results.
One of the basic principles of any type of scientific research is that the results must be replicable.
Replication means repeating an experiment in the same way to produce the same results. By clearly detailing the specifics of how the variables were measured and manipulated, other researchers can better understand the results and repeat the study if needed.
Some variables are more difficult than others to define. For example, how would you operationally define a variable such as aggression ? For obvious ethical reasons, researchers cannot create a situation in which a person behaves aggressively toward others.
To measure this variable, the researcher must devise a measurement that assesses aggressive behavior without harming others. The researcher might utilize a simulated task to measure aggressiveness in this situation.
The hypothesis you use will depend on what you are investigating and hoping to find. Some of the main types of hypotheses that you might use include:
A hypothesis often follows a basic format of "If {this happens} then {this will happen}." One way to structure your hypothesis is to describe what will happen to the dependent variable if you change the independent variable .
The basic format might be: "If {these changes are made to a certain independent variable}, then we will observe {a change in a specific dependent variable}."
Once a researcher has formed a testable hypothesis, the next step is to select a research design and start collecting data. The research method depends largely on exactly what they are studying. There are two basic types of research methods: descriptive research and experimental research.
Descriptive research such as case studies , naturalistic observations , and surveys are often used when conducting an experiment is difficult or impossible. These methods are best used to describe different aspects of a behavior or psychological phenomenon.
Once a researcher has collected data using descriptive methods, a correlational study can examine how the variables are related. This research method might be used to investigate a hypothesis that is difficult to test experimentally.
Experimental methods are used to demonstrate causal relationships between variables. In an experiment, the researcher systematically manipulates a variable of interest (known as the independent variable) and measures the effect on another variable (known as the dependent variable).
Unlike correlational studies, which can only be used to determine if there is a relationship between two variables, experimental methods can be used to determine the actual nature of the relationship—whether changes in one variable actually cause another to change.
The hypothesis is a critical part of any scientific exploration. It represents what researchers expect to find in a study or experiment. In situations where the hypothesis is unsupported by the research, the research still has value. Such research helps us better understand how different aspects of the natural world relate to one another. It also helps us develop new hypotheses that can then be tested in the future.
Thompson WH, Skau S. On the scope of scientific hypotheses . R Soc Open Sci . 2023;10(8):230607. doi:10.1098/rsos.230607
Taran S, Adhikari NKJ, Fan E. Falsifiability in medicine: what clinicians can learn from Karl Popper [published correction appears in Intensive Care Med. 2021 Jun 17;:]. Intensive Care Med . 2021;47(9):1054-1056. doi:10.1007/s00134-021-06432-z
Eyler AA. Research Methods for Public Health . 1st ed. Springer Publishing Company; 2020. doi:10.1891/9780826182067.0004
Nosek BA, Errington TM. What is replication ? PLoS Biol . 2020;18(3):e3000691. doi:10.1371/journal.pbio.3000691
Aggarwal R, Ranganathan P. Study designs: Part 2 - Descriptive studies . Perspect Clin Res . 2019;10(1):34-36. doi:10.4103/picr.PICR_154_18
Nevid J. Psychology: Concepts and Applications. Wadworth, 2013.
By Kendra Cherry, MSEd Kendra Cherry, MS, is a psychosocial rehabilitation specialist, psychology educator, and author of the "Everything Psychology Book."
An official website of the United States government
The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.
The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.
The PMC website is updating on October 15, 2024. Learn More or Try it out now .
Chittaranjan andrade.
1 Dept. of Clinical Psychopharmacology and Neurotoxicology, National Institute of Mental Health and Neurosciences, Bengaluru, Karnataka, India.
Students without prior research experience may not know how to conceptualize and design a study. This article explains how an understanding of the classification and operationalization of variables is the key to the process. Variables describe aspects of the sample that is under study; they are so called because they vary in value from subject to subject in the sample. Variables may be independent or dependent. Independent variables influence the value of other variables; dependent variables are influenced in value by other variables. A hypothesis states an expected relationship between variables. A significant relationship between an independent and dependent variable does not prove cause and effect; the relationship may partly or wholly be explained by one or more confounding variables. Variables need to be operationalized; that is, defined in a way that permits their accurate measurement. These and other concepts are explained with the help of clinically relevant examples.
This article explains the following concepts: Independent variables, dependent variables, confounding variables, operationalization of variables, and construction of hypotheses.
In any body of research, the subject of study requires to be described and understood. For example, if we wish to study predictors of response to antidepressant drugs (ADs) in patients with major depressive disorder (MDD), we might select patient age, sex, age at onset of MDD, number of previous episodes of depression, duration of current depressive episode, presence of psychotic symptoms, past history of response to ADs, and other patient and illness characteristics as potential predictors. These characteristics or descriptors are called variables. Whether or not the patient responds to AD treatment is also a variable. A solid understanding of variables is the cornerstone in the conceptualization and preparation of a research protocol, and in the framing of study hypotheses. This subject is presented in two parts. This article, Part 1, explains what independent and dependent variables are, how an understanding of these is important in framing hypotheses, and what operationalization of a variable entails.
Variables are defined as characteristics of the sample that are examined, measured, described, and interpreted. Variables are so called because they vary in value from subject to subject in the study. As an example, if we wish to examine the relationship between age and height in a sample of children, age and height are the variables of interest; their values vary from child to child. In the earlier example, patients vary in age, sex, duration of current depressive episode, and response to ADs. Variables are classified as dependent and independent variables and are usually analyzed as categorical or continuous variables.
Independent variables are defined as those the values of which influence other variables. For example, age, sex, current smoking, LDL cholesterol level, and blood pressure are independent variables because their values (e.g., greater age, positive for current smoking, and higher LDL cholesterol level) influence the risk of myocardial infarction. Dependent variables are defined as those the values of which are influenced by other variables. For example, the risk of myocardial infarction is a dependent variable the value of which is influenced by variables such as age, sex, current smoking, LDL cholesterol level, and blood pressure. The risk is higher in older persons, in men, in current smokers, and so on.
There may be a cause–effect relationship between independent and dependent variables. For example, consider a clinical trial with treatment (iron supplement vs placebo) as the independent variable and hemoglobin level as the dependent variable. In children with anemia, an iron supplement will raise the hemoglobin level to a greater extent than will placebo; this is a cause–effect relationship because iron is necessary for the synthesis of hemoglobin. However, consider the variables teeth and weight . An alien from outer space who has no knowledge of human physiology may study human children below the age of 5 years and find that, as the number of teeth increases, weight increases. Should the alien conclude that there is a cause–effect relationship here, and that growing teeth causes weight gain? No, because a third variable, age, is a confounding variable 1 – 3 that is responsible for both increase in the number of teeth and increase in weight. In general, therefore, it is more proper to state that independent variables are associated with variations in the values of the dependent variables rather than state that independent variables cause variations in the values of the dependent variables. For causality to be asserted, other criteria must be fulfilled; this is out of the scope of the present article, and interested readers may refer to Schunemann et al. 4
As a side note, here, whether a particular variable is independent or dependent will depend on the question that is being asked. For example, in a study of factors influencing patient satisfaction with outpatient department (OPD) services, patient satisfaction is the dependent variable. But, in a study of factors influencing OPD attendance at a hospital, OPD attendance is the dependent variable, and patient satisfaction is merely one of many possible independent variables that can influence OPD attendance.
Students must have a clear idea about what they want to study in order to conceptualize and frame a research protocol. The first matters that they need to address are “What are my research questions?” and “What are my hypotheses?” Both questions can be answered only after choosing the dependent variables and then the independent variables for study.
In the case of a student who is interested in studying predictors of AD outcomes in patients with MDD, treatment response is the dependent variable and patient and clinical characteristics are possible independent variables. So, the selection of dependent and independent variables helps defines the objectives of the study:
Note that in a formal research protocol, the student will need to state all the independent variables and not merely list examples. The student may also choose to include additional independent variables, such as baseline biochemical, psychophysiological, and neuroradiological measures.
A hypothesis is a clear statement of what the researcher expects to find in the study. As an example, a researcher may hypothesize that longer duration of current depression is associated with poorer response to ADs. In this hypothesis, the duration of the current episode of depression is the independent variable and treatment response is the dependent variable. It should be obvious, now, that a hypothesis can also be defined as the statement of an expected relationship between an independent and a dependent variable . Or, expressed visually, (independent variable) (arrow) (dependent variable) = hypothesis.
It would be a waste of time and energy to do a study to examine only one question: whether duration of current depression predicts treatment response. So, it is usual for research protocols to include many independent variables and many dependent variables in the generation of many hypotheses, as shown in Table 1 . Pairing each variable in the “independent variable” column with each variable in the “dependent variable” column would result in the generation of these hypotheses. Table 2 shows how this is done for age. Sets of hypotheses can likewise be constructed for the remaining independent and dependent variables in Table 1 . Importantly, the student must select one of these hypotheses as the primary hypothesis; the remaining hypotheses, no matter how many they are, would be secondary hypotheses. It is necessary to have only one hypothesis as the primary hypothesis in order to calculate the sample size necessary for an adequately powered study and to reduce the risk of false positive findings in the analysis. 5 In rare situations, two hypotheses may be considered equally important and may be stated as coprimary hypotheses.
Independent Variables and Dependent Variables in a Study on Sociodemographic and Clinical Prediction of Response of Major Depressive Disorder to Antidepressant Drug Treatment
• Age • Sex • Age at onset of major depressive disorder • Number of past episodes of depression • Past history of response to antidepressant drugs • Duration of current depressive episode • Baseline severity of depression • Baseline suicidality • Baseline melancholia • Baseline psychotic symptoms • Baseline soft neurological signs • Severity of depression • Global severity of illness • Subjective well-being • Quality of life • Everyday functioning |
Combinations of Age with Dependent Variables in the Generation of Hypotheses
1. Older age is associated with less attenuation in the severity of depression. 2. Older age is associated with less attenuation in the global severity of illness. 3. Older age is associated with less improvement in subjective well-being. 4. Older age is associated with less improvement in quality of life. 5. Older age is associated with less improvement in everyday functioning. |
In Table 1 , suicidality is listed as an independent variable and severity of depression, as a dependent variable. These variables need to be operationalized; that is, stated in a way that explains how they will be measured. Table 3 presents three ways in which suicidality can be measured and four ways in which (reduction in) the severity of depression can be measured. Now, each way of measurement in the “independent variable” column can be paired with a way of measurement in the “dependent variable” column, making a total of 12 possible hypotheses. In like manner, the many variables listed in Table 1 can each be operationalized in several different ways, resulting in the generation of a very large number of hypotheses. As already stated, the student must select only one hypothesis as the primary hypothesis.
Possible Ways of Operationalization of Suicidality and Depression
Independent Variable: Suicidality | Dependent Variable: Severity of Depression |
• Item score on the HAM-D • Item score on the MADRS • Beck scale for Suicide ideation total score | • MADRS total score • HAM-D total score • HAM-D response rate • HAM-D remission rate |
HAM-D: Hamilton Depression Rating Scale, MADRS: Montgomery–Asberg Depression Rating Scale.
Much thought should be given to the operationalization of variables because variables that are carelessly operationalized will be poorly measured; the data collected will then be of poor quality, and the study will yield unreliable results. For example, socioeconomic status may be operationalized as lower, middle, or upper class, depending on the patient’s monthly income, on the total monthly income of the family, or using a validated socioeconomic status assessment scale that takes into consideration income, education, occupation, and place of residence. The student must choose the method that would best suit the needs of the study, and the method that has the greatest scientific acceptability. However, it is also permissible to operationalize the same variable in many different ways and to include all these different operationalizations in the study, as shown in Table 3 . This is because conceptualizing variables in different ways can help understand the subject of the study in different ways.
Operationalization of variables requires a consideration of the reliability and validity of the method of operationalization; discussions on reliability and validity are out of the scope of this article. Operationalization of variables also requires specification of the scale of measurement: nominal, ordinal, interval, or ratio; this is also out of the scope of the present article. Finally, operationalization of variables can also specify details of the measurement procedure. As an example, in a study on the use of metformin to reduce olanzapine-associated weight gain, we may state that we will obtain the weight of the patient but fail to explain how we will do it. Better would be to state that the same weighing scale will be used. Still better would be to state that we will use a weighing instrument that works on the principle of moving weights on a levered arm, and that the same instrument will be used for all patients. And best would be to add that we will weigh patients, dressed in standard hospital gowns, after they have voided their bladder but before they have eaten breakfast. When the way in which a variable will be measured is defined, measurement of that variable becomes more objective and uniform
The next article, Part 2, will address what categorical and continuous variables are, why continuous variables should not be converted into categorical variables and when this rule can be broken, and what confounding variables are.
Declaration of Conflicting Interests: The author declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding: The author received no financial support for the research, authorship, and/or publication of this article.
Independent and dependent variables in research, can qualitative data have independent and dependent variables.
Experiments rely on capturing the relationship between independent and dependent variables to understand causal patterns. Researchers can observe what happens when they change a condition in their experiment or if there is any effect at all.
It's important to understand the difference between the independent variable and dependent variable. We'll look at the notion of independent and dependent variables in this article. If you are conducting experimental research, defining the variables in your study is essential for realizing rigorous research .
In experimental research, a variable refers to the phenomenon, person, or thing that is being measured and observed by the researcher. A researcher conducts a study to see how one variable affects another and make assertions about the relationship between different variables.
A typical research question in an experimental study addresses a hypothesized relationship between the independent variable manipulated by the researcher and the dependent variable that is the outcome of interest presumably influenced by the researcher's manipulation.
Take a simple experiment on plants as an example. Suppose you have a control group of plants on one side of a garden and an experimental group of plants on the other side. All things such as sunlight, water, and fertilizer being equal, both plants should be expected to grow at the same rate.
Now imagine that the plants in the experimental group are given a new plant fertilizer under the assumption that they will grow faster. Then you will need to measure the difference in growth between the two groups in your study.
In this case, the independent variable is the type of fertilizer used on your plants while the dependent variable is the rate of growth among your plants. If there is a significant difference in growth between the two groups, then your study provides support to suggest that the fertilizer causes higher rates of plant growth.
The independent variable is the element in your study that you intentionally change, which is why it can also be referred to as the manipulated variable.
You manipulate this variable to see how it might affect the other variables you observe, all other factors being equal. This means that you can observe the cause and effect relationships between one independent variable and one or multiple dependent variables.
Independent variables are directly manipulated by the researcher, while dependent variables are not. They are "dependent" because they are affected by the independent variable in the experiment. Researchers can thus study how manipulating the independent variable leads to changes in the main outcome of interest being measured as the dependent variable.
Note that while you can have multiple dependent variables, it is challenging to establish research rigor for multiple independent variables. If you are making so many changes in an experiment, how do you know which change is responsible for the outcome produced by the study? Studying more than one independent variable would require running an experiment for each independent variable to isolate its effects on the dependent variable.
This being said, it is certainly possible to employ a study design that involves multiple independent and dependent variables, as is the case with what is called a factorial experiment. For example, a psychological study examining the effects of sleep and stress levels on work productivity and social interaction would have two independent variables and two dependent variables, respectively.
Such a study would be complex and require careful planning to establish the necessary research rigor , however. If possible, consider narrowing your research to the examination of one independent variable to make it more manageable and easier to understand.
Let's consider an experiment in the social studies. Suppose you want to determine the effectiveness of a new textbook compared to current textbooks in a particular school.
The new textbook is supposed to be better, but how can you prove it? Besides all the selling points that the textbook publisher makes, how do you know if the new textbook is any good? A rigorous study examining the effects of the textbook on classroom outcomes is in order.
The textbook given to students makes up the independent variable in your experimental study. The shift from the existing textbooks to the new one represents the manipulation of the independent variable in this study.
In any experiment, the dependent variable is observed to measure how it is affected by changes to the independent variable. Outcomes such as test scores and other performance metrics can make up the data for the dependent variable.
Now that we are changing the textbook in the experiment above, we should examine if there are any effects.
To do this, we will need two classrooms of students. As best as possible, the two sets of students should be of similar proficiency (or at least of similar backgrounds) and placed within similar conditions for teaching and learning (e.g., physical space, lesson planning).
The control group in our study will be one set of students using the existing textbook. By examining their performance, we can establish a baseline. The performance of the experimental group, which is the set of students using the new textbook, can then be compared with the baseline performance.
As a result, the change in the test scores make up the data for our dependent variable. We cannot directly affect how well students perform on the test, but we can conclude from our experiment whether the use of the new textbook might impact students' performance.
Rely on our powerful data analysis interface for your research, starting with a free trial.
We can typically think of an independent variable as something a researcher can directly change. In the above example, we can change the textbook used by the teacher in class. If we're talking about plants, we can change the fertilizer.
Conversely, the dependent variable is something that we do not directly influence or manipulate. Strictly speaking, we cannot directly manipulate a student's performance on a test or the rate of growth of a plant, not without other factors such as new teaching methods or new fertilizer, respectively.
Understanding the distinction between a dependent variable and an independent variable is key to experimental research. Ultimately, the distinction can be reduced to which element in a study has been directly influenced by the researcher.
Given the potential complexities encountered in research, there is essential terminology for other variables in any experimental study. You might employ this terminology or encounter them while reading other research.
A control variable is any factor that the researcher tries to keep constant as the independent variable changes. In the plant experiment described earlier in this article, the sunlight and water are each a controlled variable while the type of fertilizer used is the manipulated variable across control and experimental groups.
To ensure research rigor, the researcher needs to keep these control variables constant to dispel any concerns that differences in growth rate were being driven by sunlight or water, as opposed to the fertilizer being used.
Extraneous variables refer to any unwanted influence on the dependent variable that may confound the analysis of the study. For example, if bugs or animals ate the plants in your fertilizer study, this was greatly impact the rates of plant growth. This is why it would be important to control the environment and protect it from such threats.
Finally, independent variables can go by different names such as subject variables or predictor variables. Dependent variables can also be referred to as the responding variable or outcome variable. Whatever the language, they all serve the same role of influencing the dependent variable in an experiment.
The use of the word " variables " is typically associated with quantitative and confirmatory research. Naturalistic qualitative research typically does not employ experimental designs or establish causality. Qualitative research often draws on observations , interviews , focus groups , and other forms of data collection that are allow researchers to study the naturally occurring "messiness" of the social world, rather than controlling all variables to isolate a cause-and-effect relationship.
In limited circumstances, the idea of experimental variables can apply to participant observations in ethnography , where the researcher should be mindful of their influence on the environment they are observing.
However, the experimental paradigm is best left to quantitative studies and confirmatory research questions. Qualitative researchers in the social sciences are oftentimes more interested in observing and describing socially-constructed phenomena rather than testing hypotheses .
Nonetheless, the notion of independent and dependent variables does hold important lessons for qualitative researchers. Even if they don't employ variables in their study design, qualitative researchers often observe how one thing affects another. A theoretical or conceptual framework can then suggest potential cause-and-effect relationships in their study.
Download a free trial of ATLAS.ti to see how you can make the most of your data.
Run a free plagiarism check in 10 minutes, generate accurate citations for free.
Published on May 6, 2022 by Shaun Turney . Revised on June 22, 2023.
The null and alternative hypotheses are two competing claims that researchers weigh evidence for and against using a statistical test :
Answering your research question with hypotheses, what is a null hypothesis, what is an alternative hypothesis, similarities and differences between null and alternative hypotheses, how to write null and alternative hypotheses, other interesting articles, frequently asked questions.
The null and alternative hypotheses offer competing answers to your research question . When the research question asks “Does the independent variable affect the dependent variable?”:
The null and alternative are always claims about the population. That’s because the goal of hypothesis testing is to make inferences about a population based on a sample . Often, we infer whether there’s an effect in the population by looking at differences between groups or relationships between variables in the sample. It’s critical for your research to write strong hypotheses .
You can use a statistical test to decide whether the evidence favors the null or alternative hypothesis. Each type of statistical test comes with a specific way of phrasing the null and alternative hypothesis. However, the hypotheses can also be phrased in a general way that applies to any test.
Professional editors proofread and edit your paper by focusing on:
See an example
The null hypothesis is the claim that there’s no effect in the population.
If the sample provides enough evidence against the claim that there’s no effect in the population ( p ≤ α), then we can reject the null hypothesis . Otherwise, we fail to reject the null hypothesis.
Although “fail to reject” may sound awkward, it’s the only wording that statisticians accept . Be careful not to say you “prove” or “accept” the null hypothesis.
Null hypotheses often include phrases such as “no effect,” “no difference,” or “no relationship.” When written in mathematical terms, they always include an equality (usually =, but sometimes ≥ or ≤).
You can never know with complete certainty whether there is an effect in the population. Some percentage of the time, your inference about the population will be incorrect. When you incorrectly reject the null hypothesis, it’s called a type I error . When you incorrectly fail to reject it, it’s a type II error.
The table below gives examples of research questions and null hypotheses. There’s always more than one way to answer a research question, but these null hypotheses can help you get started.
( ) | ||
Does tooth flossing affect the number of cavities? | Tooth flossing has on the number of cavities. | test: The mean number of cavities per person does not differ between the flossing group (µ ) and the non-flossing group (µ ) in the population; µ = µ . |
Does the amount of text highlighted in the textbook affect exam scores? | The amount of text highlighted in the textbook has on exam scores. | : There is no relationship between the amount of text highlighted and exam scores in the population; β = 0. |
Does daily meditation decrease the incidence of depression? | Daily meditation the incidence of depression.* | test: The proportion of people with depression in the daily-meditation group ( ) is greater than or equal to the no-meditation group ( ) in the population; ≥ . |
*Note that some researchers prefer to always write the null hypothesis in terms of “no effect” and “=”. It would be fine to say that daily meditation has no effect on the incidence of depression and p 1 = p 2 .
The alternative hypothesis ( H a ) is the other answer to your research question . It claims that there’s an effect in the population.
Often, your alternative hypothesis is the same as your research hypothesis. In other words, it’s the claim that you expect or hope will be true.
The alternative hypothesis is the complement to the null hypothesis. Null and alternative hypotheses are exhaustive, meaning that together they cover every possible outcome. They are also mutually exclusive, meaning that only one can be true at a time.
Alternative hypotheses often include phrases such as “an effect,” “a difference,” or “a relationship.” When alternative hypotheses are written in mathematical terms, they always include an inequality (usually ≠, but sometimes < or >). As with null hypotheses, there are many acceptable ways to phrase an alternative hypothesis.
The table below gives examples of research questions and alternative hypotheses to help you get started with formulating your own.
Does tooth flossing affect the number of cavities? | Tooth flossing has an on the number of cavities. | test: The mean number of cavities per person differs between the flossing group (µ ) and the non-flossing group (µ ) in the population; µ ≠ µ . |
Does the amount of text highlighted in a textbook affect exam scores? | The amount of text highlighted in the textbook has an on exam scores. | : There is a relationship between the amount of text highlighted and exam scores in the population; β ≠ 0. |
Does daily meditation decrease the incidence of depression? | Daily meditation the incidence of depression. | test: The proportion of people with depression in the daily-meditation group ( ) is less than the no-meditation group ( ) in the population; < . |
Null and alternative hypotheses are similar in some ways:
However, there are important differences between the two types of hypotheses, summarized in the following table.
A claim that there is in the population. | A claim that there is in the population. | |
| ||
Equality symbol (=, ≥, or ≤) | Inequality symbol (≠, <, or >) | |
Rejected | Supported | |
Failed to reject | Not supported |
Discover proofreading & editing
To help you write your hypotheses, you can use the template sentences below. If you know which statistical test you’re going to use, you can use the test-specific template sentences. Otherwise, you can use the general template sentences.
The only thing you need to know to use these general template sentences are your dependent and independent variables. To write your research question, null hypothesis, and alternative hypothesis, fill in the following sentences with your variables:
Does independent variable affect dependent variable ?
Once you know the statistical test you’ll be using, you can write your hypotheses in a more precise and mathematical way specific to the test you chose. The table below provides template sentences for common statistical tests.
( ) | ||
test
with two groups | The mean dependent variable does not differ between group 1 (µ ) and group 2 (µ ) in the population; µ = µ . | The mean dependent variable differs between group 1 (µ ) and group 2 (µ ) in the population; µ ≠ µ . |
with three groups | The mean dependent variable does not differ between group 1 (µ ), group 2 (µ ), and group 3 (µ ) in the population; µ = µ = µ . | The mean dependent variable of group 1 (µ ), group 2 (µ ), and group 3 (µ ) are not all equal in the population. |
There is no correlation between independent variable and dependent variable in the population; ρ = 0. | There is a correlation between independent variable and dependent variable in the population; ρ ≠ 0. | |
There is no relationship between independent variable and dependent variable in the population; β = 0. | There is a relationship between independent variable and dependent variable in the population; β ≠ 0. | |
Two-proportions test | The dependent variable expressed as a proportion does not differ between group 1 ( ) and group 2 ( ) in the population; = . | The dependent variable expressed as a proportion differs between group 1 ( ) and group 2 ( ) in the population; ≠ . |
Note: The template sentences above assume that you’re performing one-tailed tests . One-tailed tests are appropriate for most studies.
If you want to know more about statistics , methodology , or research bias , make sure to check out some of our other articles with explanations and examples.
Methodology
Research bias
Hypothesis testing is a formal procedure for investigating our ideas about the world using statistics. It is used by scientists to test specific predictions, called hypotheses , by calculating how likely it is that a pattern or relationship between variables could have arisen by chance.
Null and alternative hypotheses are used in statistical hypothesis testing . The null hypothesis of a test always predicts no effect or no relationship between variables, while the alternative hypothesis states your research prediction of an effect or relationship.
The null hypothesis is often abbreviated as H 0 . When the null hypothesis is written using mathematical symbols, it always includes an equality symbol (usually =, but sometimes ≥ or ≤).
The alternative hypothesis is often abbreviated as H a or H 1 . When the alternative hypothesis is written using mathematical symbols, it always includes an inequality symbol (usually ≠, but sometimes < or >).
A research hypothesis is your proposed answer to your research question. The research hypothesis usually includes an explanation (“ x affects y because …”).
A statistical hypothesis, on the other hand, is a mathematical statement about a population parameter. Statistical hypotheses always come in pairs: the null and alternative hypotheses . In a well-designed study , the statistical hypotheses correspond logically to the research hypothesis.
If you want to cite this source, you can copy and paste the citation or click the “Cite this Scribbr article” button to automatically add the citation to our free Citation Generator.
Turney, S. (2023, June 22). Null & Alternative Hypotheses | Definitions, Templates & Examples. Scribbr. Retrieved September 16, 2024, from https://www.scribbr.com/statistics/null-and-alternative-hypotheses/
Other students also liked, inferential statistics | an easy introduction & examples, hypothesis testing | a step-by-step guide with easy examples, type i & type ii errors | differences, examples, visualizations, what is your plagiarism score.
As you were browsing something about your browser made us think you were a bot. There are a few reasons this might happen:
To regain access, please make sure that cookies and JavaScript are enabled before reloading the page.
IMAGES
VIDEO
COMMENTS
Examples of Independent and Dependent Variables. 1. Gatorade and Improved Athletic Performance. A sports medicine researcher has been hired by Gatorade to test the effects of its sports drink on athletic performance. The company wants to claim that when an athlete drinks Gatorade, their performance will improve.
The independent variable is the cause. Its value is independent of other variables in your study. The dependent variable is the effect. Its value depends on changes in the independent variable. Example: Independent and dependent variables. You design a study to test whether changes in room temperature have an effect on math test scores.
Here are several examples of independent and dependent variables in experiments: In a study to determine whether how long a student sleeps affects test scores, the independent variable is the length of time spent sleeping while the dependent variable is the test score. You want to know which brand of fertilizer is best for your plants.
Independent and Dependent Variable Examples. In a study to determine whether the amount of time a student sleeps affects test scores, the independent variable is the amount of time spent sleeping while the dependent variable is the test score. You want to compare brands of paper towels to see which holds the most liquid.
While the independent variable is the " cause ", the dependent variable is the " effect " - or rather, the affected variable. In other words, the dependent variable is the variable that is assumed to change as a result of a change in the independent variable. Keeping with the previous example, let's look at some dependent variables ...
5. Phrase your hypothesis in three ways. To identify the variables, you can write a simple prediction in if…then form. The first part of the sentence states the independent variable and the second part states the dependent variable. If a first-year student starts attending more lectures, then their exam scores will improve.
In research, the independent variable is manipulated to observe its effect, while the dependent variable is the measured outcome. Essentially, the independent variable is the presumed cause, and the dependent variable is the observed effect. Variables provide the foundation for examining relationships, drawing conclusions, and making ...
The researcher manipulates the independent variable and the dependent variable is the measured outcome. Operationalized the variables being investigated. Operationalization of a hypothesis refers to the process of making the variables physically measurable or testable, e.g. if you are about to study aggression, you might count the number of ...
The independent variable is the cause. Its value is independent of other variables in your study. The dependent variable is the effect. Its value depends on changes in the independent variable. Example: Independent and dependent variables. You design a study to test whether changes in room temperature have an effect on maths test scores.
A dependent variable is something the researcher observes and measures. Example: Hypothesis Daily exposure to the sun leads to increased levels of happiness. In this example, the independent variable is exposure to the sun - the assumed cause. The dependent variable is the level of happiness - the assumed effect.
Independent and Dependent Variables: Differences & Examples. By Jim Frost 15 Comments. Independent variables and dependent variables are the two fundamental types of variables in statistical modeling and experimental designs. Analysts use these methods to understand the relationships between the variables and estimate effect sizes.
A simple hypothesis suggests only the relationship between two variables: one independent and one dependent. Examples: If you stay up late, then you feel tired the next day. Turning off your phone makes it charge faster. 2 Complex hypothesis. A complex hypothesis suggests the relationship between more than two variables, for example, two ...
Simple hypothesis: This type of hypothesis suggests there is a relationship between one independent variable and one dependent variable.; Complex hypothesis: This type suggests a relationship between three or more variables, such as two independent and dependent variables.; Null hypothesis: This hypothesis suggests no relationship exists between two or more variables.
The dependent variable is the factor which the researcher hypothesizes will change in response to the independent variable; it is the measured outcome of the experiment.
Dependent variable: blood pressure. Example 3: A researcher changes the version of a study guide given to students to see how it affects exam scores. Independent variable: the version of the study guide. Dependent variable: exam scores. Independent vs. Dependent Variables on a Graph. When we create a graph, the independent variable will go on ...
DEFINING VARIABLES WITHIN A HYPOTHESIS Types of Variables: Dependent Variable: variable of interest, acted upon by the independent variable. Independent Variable: variable suspected to control/influence the dependent variable. Controlled Variable: variable(s) not included in the hypothesis that could influence the dependent variable and should be held constant in an experiment.
A variable is considered dependent if it depends on an independent variable. Dependent variables are studied under the supposition or ... and goodness of fit, but are not of substantive interest to the hypothesis under examination. For example, in a study examining the effect of post-secondary education on lifetime earnings, some extraneous ...
There may be a cause-effect relationship between independent and dependent variables. For example, consider a clinical trial with treatment (iron supplement vs placebo) as the independent variable and hemoglobin level as the dependent variable. In children with anemia, an iron supplement will raise the hemoglobin level to a greater extent ...
Variables are an important concept in experimental and hypothesis-testing research, so understanding independent/dependent variables is key to understanding research design. In this article, we will talk about what separates a dependent variable from an independent variable and how the concept applies to research.
The mean dependent variable of group 1 (µ 1), group 2 (µ 2), and group 3 (µ 3) are not all equal in the population. Pearson correlation: There is no correlation between independent variable and dependent variable in the population; ρ = 0. There is a correlation between independent variable and dependent variable in the population; ρ ≠ 0.
Independent Variables. The independent variable is the variable that the researcher or experimenter manipulates to affect the dependent variable. It is independent of the other variables in an experiment. In other words, the independent variable causes some kind of change in the dependent variable.
For those that do have a hypothesis, the hypothesis should derive logically from previous findings or the predictions of a particular theory. Hypotheses should not be based simply on what the student believes should happen. A clear rationale is necessary. An examination of publications by student researchers provides several examples of hypotheses.
To form a solid theory, the vital first step is creating a hypothesis. See the various types of hypotheses and how they can lead you on the path to discovery.
Sam Levasseur SOSC 341: Independent and Dependent Variables in Research Articles Professor Murphy 08/28/2024 Article 1: Gamble, T., & Walker, I. (2016). Wearing a bicycle helmet can increase risk- taking and sensation-seeking in adults. The study found that adults who wore bicycle helmets exhibited higher risk-taking and sensation-seeking behaviors than those who did not, highlighting the ...