Directional vs Non-Directional Hypothesis: Key Difference

In statistics, a directional hypothesis, also known as a one-tailed hypothesis, is a type of hypothesis that predicts the direction of the relationship between variables or the direction of the difference between groups.

non directional two tailed hypothesis

The introduction of a directional hypothesis in a research study provides an overview of the specific prediction being made about the relationship between variables or the difference between groups. It sets the stage for the research question and outlines the expected direction of the findings. The introduction typically includes the following elements:

Research Context: Begin by introducing the general topic or research area that the study is focused on. Provide background information and highlight the significance of the research question.

Research Question: Clearly state the specific research question that the study aims to answer. This question should be directly related to the variables being investigated.

Previous Research: Summarize relevant literature or previous studies that have explored similar or related topics. This helps establish the existing knowledge base and provides a rationale for the hypothesis.

Hypothesis Statement: Present the directional hypothesis clearly and concisely. State the predicted relationship between variables or the expected difference between groups. For example, if studying the impact of a new teaching method on student performance, a directional hypothesis could be, “Students who receive the new teaching method will demonstrate higher test scores compared to students who receive the traditional teaching method.”

Justification: Provide a logical explanation for the directional hypothesis based on the existing literature or theoretical framework . Discuss any previous findings, theories, or empirical evidence that support the predicted direction of the relationship or difference.

Objectives: Outline the specific objectives or aims of the study, which should align with the research question and hypothesis. These objectives help guide the research process and provide a clear focus for the study.

By including these elements in the introduction of a research study, the directional hypothesis is introduced effectively, providing a clear and justified prediction about the expected outcome of the research.

When formulating a directional hypothesis, researchers make a specific prediction about the expected relationship or difference between variables. They specify whether they expect an increase or decrease in the dependent variable, or whether one group will score higher or lower than another group

What is Directional Hypothesis?

With a correlational study, a directional hypothesis states that there is a positive (or negative) correlation between two variables. When a hypothesis states the direction of the results, it is referred to as a directional (one-tailed) hypothesis; this is because it states that the results go in one direction.

Definition:

A directional hypothesis is a one-tailed hypothesis that states the direction of the difference or relationship (e.g. boys are more helpful than girls).

Research Question: Does exercise have a positive impact on mood?

Directional Hypothesis: Engaging in regular exercise will result in an increase in positive mood compared to a sedentary lifestyle.

In this example, the directional hypothesis predicts that regular exercise will have a specific effect on mood, specifically leading to an increase in positive mood. The researcher expects that individuals who engage in regular exercise will experience improvements in their overall mood compared to individuals who lead a sedentary lifestyle.

It’s important to note that this is just one example, and directional hypotheses can be formulated in various research areas and contexts. The key is to make a specific prediction about the direction of the relationship or difference between variables based on prior knowledge or theoretical considerations.

Advantages of Directional Hypothesis

There are several advantages to using a directional hypothesis in research studies. Here are a few key benefits:

Specific Prediction:

A directional hypothesis allows researchers to make a specific prediction about the expected relationship or difference between variables. This provides a clear focus for the study and helps guide the research process. It also allows for more precise interpretation of the results.

Testable and Refutable:

Directional hypotheses can be tested and either supported or refuted by empirical evidence. Researchers can design their study and select appropriate statistical tests to specifically examine the predicted direction of the relationship or difference. This enhances the rigor and validity of the research.

Efficiency and Resource Allocation:

By making a specific prediction, researchers can allocate their resources more efficiently. They can focus on collecting data and conducting analyses that directly test the directional hypothesis, rather than exploring all possible directions or relationships. This can save time, effort, and resources.

Theory Development:

Directional hypotheses contribute to the development of theories and scientific knowledge. When a directional hypothesis is supported by empirical evidence, it provides support for existing theories or helps generate new theories. This advancement in knowledge can guide future research and understanding in the field.

Practical Applications:

Directional hypotheses can have practical implications and applications. If a hypothesis predicts a specific direction of change, such as the effectiveness of a treatment or intervention, it can inform decision-making and guide practical applications in fields such as medicine, psychology, or education.

Enhanced Communication:

Directional hypotheses facilitate clearer communication of research findings. When researchers have made specific predictions about the direction of the relationship or difference, they can effectively communicate their results to both academic and non-academic audiences. This promotes better understanding and application of the research outcomes.

It’s important to note that while directional hypotheses offer advantages, they also require stronger evidence to support them compared to non-directional hypotheses. Researchers should carefully consider the research context, existing literature, and theoretical considerations before formulating a directional hypothesis.

Disadvantages of Directional Hypothesis

While directional hypotheses have their advantages, there are also some potential disadvantages to consider:

Risk of Type I Error:

Directional hypotheses increase the risk of committing a Type I error , also known as a false positive. By focusing on a specific predicted direction, researchers may overlook the possibility of an opposite or null effect. If the actual relationship or difference does not align with the predicted direction, researchers may incorrectly conclude that there is no effect when, in fact, there may be.

Narrow Focus:

Directional hypotheses restrict the scope of investigation to a specific predicted direction. This narrow focus may overlook other potential relationships, nuances, or alternative explanations. Researchers may miss valuable insights or unexpected findings by excluding other possibilities from consideration.

Limited Generalizability:

Directional hypotheses may limit the generalizability of findings. If the study supports the predicted direction, the results may only apply to the specific context and conditions outlined in the hypothesis. Generalizing the findings to different populations, settings, or variables may require further research.

Biased Interpretation:

Directional hypotheses can introduce bias in the interpretation of results. Researchers may be inclined to selectively focus on evidence that supports the predicted direction while downplaying or ignoring contradictory evidence. This can hinder objectivity and lead to biased conclusions.

Increased Sample Size Requirements:

Directional hypotheses often require larger sample sizes compared to non-directional hypotheses. This is because statistical power needs to be sufficient to detect the predicted direction with a reasonable level of confidence. Larger samples can be more time-consuming and resource-intensive to obtain.

Reduced Flexibility:

Directional hypotheses limit flexibility in data analysis and statistical testing. Researchers may feel compelled to use specific statistical tests or analytical approaches that align with the predicted direction, potentially overlooking alternative methods that may be more appropriate or informative.

It’s important to weigh these disadvantages against the specific research context and objectives when deciding whether to use a directional hypothesis. In some cases, a non-directional hypothesis may be more suitable, allowing for a more exploratory and comprehensive investigation of the research question.

Non-Directional Hypothesis:

A non-directional hypothesis, also known as a two-tailed hypothesis, is a type of hypothesis that does not specify the direction of the relationship between variables or the difference between groups. Instead of predicting a specific direction, a non-directional hypothesis suggests that there will be a significant relationship or difference, without indicating whether it will be positive or negative, higher or lower, etc.

The introduction of a non-directional hypothesis in a research study provides an overview of the general prediction being made about the relationship between variables or the difference between groups, without specifying the direction. It sets the stage for the research question and outlines the expectation of a significant relationship or difference. The introduction typically includes the following elements:

Research Context:

Begin by introducing the general topic or research area that the study is focused on. Provide background information and highlight the significance of the research question.

Research Question:

Clearly state the specific research question that the study aims to answer. This question should be directly related to the variables being investigated.

Previous Research:

Summarize relevant literature or previous studies that have explored similar or related topics. This helps establish the existing knowledge base and provides a rationale for the hypothesis.

Hypothesis Statement:

Present the non-directional hypothesis clearly and concisely. State that there is an expected relationship or difference between variables or groups without specifying the direction. For example, if studying the relationship between socioeconomic status and academic achievement, a non-directional hypothesis could be, “There is a significant relationship between socioeconomic status and academic achievement.”

Justification:

Provide a logical explanation for the non-directional hypothesis based on the existing literature or theoretical framework. Discuss any previous findings, theories, or empirical evidence that support the notion of a relationship or difference between the variables or groups.

Objectives:

Outline the specific objectives or aims of the study, which should align with the research question and hypothesis. These objectives help guide the research process and provide a clear focus for the study.

By including these elements in the introduction of a research study, the non-directional hypothesis is introduced effectively, indicating the expectation of a significant relationship or difference without specifying the direction

What is Non-directional hypothesis?

In a non-directional hypothesis, researchers acknowledge that there may be an effect or relationship between variables but do not make a specific prediction about the direction of that effect. This allows for a more exploratory approach to data analysis and interpretation

If a hypothesis does not state a direction but simply says that one factor affects another, or that there is an association or correlation between two variables then it is called a non-directional (two-tailed) hypothesis.

Research Question: Is there a relationship between social media usage and self-esteem ?

Non-Directional Hypothesis: There is a significant relationship between social media usage and self-esteem.

In this example, the non-directional hypothesis suggests that there is a relationship between social media usage and self-esteem without specifying whether higher social media usage is associated with higher or lower self-esteem. The hypothesis acknowledges the possibility of an effect but does not make a specific prediction about the direction of that effect.

It’s important to note that this is just one example, and non-directional hypotheses can be formulated in various research areas and contexts. The key is to indicate the expectation of a significant relationship or difference without specifying the direction, allowing for a more exploratory approach to data analysis and interpretation.

Advantages of Non-directional hypothesis

Non-directional hypotheses, also known as two-tailed hypotheses, offer several advantages in research studies. Here are some of the key advantages:

Flexibility in Data Analysis:

Non-directional hypotheses allow for flexibility in data analysis. Researchers are not constrained by a specific predicted direction and can explore the relationship or difference in various ways. This flexibility enables a more comprehensive examination of the data, considering both positive and negative associations or differences.

Objective and Open-Minded Approach:

Non-directional hypotheses promote an objective and open-minded approach to research. Researchers do not have preconceived notions about the direction of the relationship or difference, which helps mitigate biases in data interpretation. They can objectively analyze the data without being influenced by their initial expectations.

Comprehensive Understanding:

By not specifying the direction, non-directional hypotheses facilitate a comprehensive understanding of the relationship or difference being investigated. Researchers can explore and consider all possible outcomes, leading to a more nuanced interpretation of the findings. This broader perspective can provide deeper insights into the research question.

Greater Sensitivity:

Non-directional hypotheses can be more sensitive to detecting unexpected or surprising relationships or differences. Researchers are not solely focused on confirming a specific predicted direction, but rather on uncovering any significant association or difference. This increased sensitivity allows for the identification of novel patterns and relationships that may have been overlooked with a directional hypothesis.

Replication and Generalizability:

Non-directional hypotheses support replication studies and enhance the generalizability of findings. By not restricting the investigation to a specific predicted direction, the results can be more applicable to different populations, contexts, or conditions. This broader applicability strengthens the validity and reliability of the research.

Hypothesis Generation:

Non-directional hypotheses can serve as a foundation for generating new hypotheses and research questions. Significant findings without a specific predicted direction can lead to further investigations and the formulation of more focused directional hypotheses in subsequent studies.

It’s important to consider the specific research context and objectives when deciding between a directional or non-directional hypothesis. Non-directional hypotheses are particularly useful when researchers are exploring new areas or when there is limited existing knowledge about the relationship or difference being studied.

Disadvantages of Non-directional hypothesis

Non-directional hypotheses have their advantages, there are also some potential disadvantages to consider:

Lack of Specificity: Non-directional hypotheses do not provide a specific prediction about the direction of the relationship or difference between variables. This lack of specificity may limit the interpretability and practical implications of the findings. Stakeholders may desire clear guidance on the expected direction of the effect.

Non-directional hypotheses often require larger sample sizes compared to directional hypotheses. This is because statistical power needs to be sufficient to detect any significant relationship or difference, regardless of the direction. Obtaining larger samples can be more time-consuming, resource-intensive, and costly.

Reduced Precision:

By not specifying the direction, non-directional hypotheses may result in less precise findings. Researchers may obtain statistically significant results indicating a relationship or difference, but the lack of direction may hinder their ability to understand the practical implications or mechanism behind the effect.

Potential for Post-hoc Interpretation:

Non-directional hypotheses can increase the risk of post-hoc interpretation of results. Researchers may be tempted to selectively interpret and highlight only the significant findings that support their preconceived notions or expectations, leading to biased interpretations.

Limited Theoretical Guidance:

Non-directional hypotheses may lack theoretical guidance in terms of understanding the underlying mechanisms or causal pathways. Without a specific predicted direction, it can be challenging to develop a comprehensive theoretical framework to explain the relationship or difference being studied.

Potential Missed Opportunities:

Non-directional hypotheses may limit the exploration of specific directions or subgroups within the data. By not focusing on a specific direction, researchers may miss important nuances or interactions that could contribute to a deeper understanding of the phenomenon under investigation.

It’s important to carefully consider the research question, available literature, and research objectives when deciding whether to use a non-directional hypothesis. Depending on the context and goals of the study, a non-directional hypothesis may be appropriate, but researchers should also be aware of the potential limitations and address them accordingly in their research design and interpretation of results.

Difference between directional and non-directional hypothesis

the main difference between a directional hypothesis and a non-directional hypothesis lies in the specificity of the prediction made about the relationship between variables or the difference between groups.

Directional Hypothesis:

A directional hypothesis, also known as a one-tailed hypothesis, makes a specific prediction about the direction of the relationship or difference. It states the expected outcome, whether it is a positive or negative relationship, a higher or lower value, an increase or decrease, etc. The directional hypothesis guides the research in a focused manner, specifying the direction to be tested.

Example: “Students who receive tutoring will demonstrate higher test scores compared to students who do not receive tutoring.”

A non-directional hypothesis, also known as a two-tailed hypothesis, does not specify the direction of the relationship or difference. It acknowledges the possibility of a relationship or difference between variables without predicting a specific direction. The non-directional hypothesis allows for exploration and analysis of both positive and negative associations or differences.

Example: “There is a significant relationship between sleep quality and academic performance.”

In summary, a directional hypothesis makes a specific prediction about the direction of the relationship or difference, while a non-directional hypothesis suggests a relationship or difference without specifying the direction. The choice between the two depends on the research question, existing literature, and the researcher’s objectives. Directional hypotheses provide a focused prediction, while non-directional hypotheses allow for more exploratory analysis .

When to use Directional Hypothesis?

A directional hypothesis is appropriate to use in specific situations where researchers have a clear theoretical or empirical basis for predicting the direction of the relationship or difference between variables. Here are some scenarios where a directional hypothesis is commonly employed:

Prior Research and Theoretical Framework: When previous studies, existing theories, or established empirical evidence strongly suggest a specific direction of the relationship or difference, a directional hypothesis can be formulated. Researchers can build upon the existing knowledge base and make a focused prediction based on this prior information.

Cause-and-Effect Relationships: In studies aiming to establish cause-and-effect relationships, directional hypotheses are often used. When there is a clear theoretical understanding of the causal relationship between variables, researchers can predict the expected direction of the effect based on the proposed mechanism.

Specific Research Objectives: If the research study has specific objectives that require a clear prediction about the direction, a directional hypothesis can be appropriate. For instance, if the aim is to test the effectiveness of a particular intervention or treatment, a directional hypothesis can guide the evaluation by predicting the expected positive or negative outcome.

Practical Applications: Directional hypotheses are useful when the research findings have direct practical implications. For example, in fields such as medicine, psychology, or education, researchers may formulate directional hypotheses to predict the effects of certain interventions or treatments on patient outcomes or educational achievement.

Hypothesis-Testing Approach: Researchers who adopt a hypothesis-testing approach, where they aim to confirm or disconfirm specific predictions, often use directional hypotheses. This approach involves formulating a specific hypothesis and conducting statistical tests to determine whether the data support or refute the predicted direction of the relationship or difference.

When to use non directional hypothesis?

A non-directional hypothesis, also known as a two-tailed hypothesis, is appropriate to use in several situations where researchers do not have a specific prediction about the direction of the relationship or difference between variables. Here are some scenarios where a non-directional hypothesis is commonly employed:

Exploratory Research:

When the research aims to explore a new area or investigate a relationship that has limited prior research or theoretical guidance, a non-directional hypothesis is often used. It allows researchers to gather initial data and insights without being constrained by a specific predicted direction.

Preliminary Studies:

Non-directional hypotheses are useful in preliminary or pilot studies that seek to gather preliminary evidence and generate hypotheses for further investigation. By using a non-directional hypothesis, researchers can gather initial data to inform the development of more specific hypotheses in subsequent studies.

Neutral Expectations:

If researchers have no theoretical or empirical basis to predict the direction of the relationship or difference, a non-directional hypothesis is appropriate. This may occur in situations where there is a lack of prior research, conflicting findings, or inconclusive evidence to support a specific direction.

Comparative Studies:

In studies where the objective is to compare two or more groups or conditions, a non-directional hypothesis is commonly used. The focus is on determining whether a significant difference exists, without making specific predictions about which group or condition will have higher or lower values.

Data-Driven Approach:

When researchers adopt a data-driven or exploratory approach to analysis, non-directional hypotheses are preferred. Instead of testing specific predictions, the aim is to explore the data, identify patterns, and generate hypotheses based on the observed relationships or differences.

Hypothesis-Generating Studies:

Non-directional hypotheses are often used in studies aimed at generating new hypotheses and research questions. By exploring associations or differences without specifying the direction, researchers can identify potential relationships or factors that can serve as a basis for future research.

Strategies to improve directional and non-directional hypothesis

To improve the quality of both directional and non-directional hypotheses, researchers can employ various strategies. Here are some strategies to enhance the formulation of hypotheses:

Strategies to Improve Directional Hypotheses:

Review existing literature:.

Conduct a thorough review of relevant literature to identify previous research findings, theories, and empirical evidence related to the variables of interest. This will help inform and support the formulation of a specific directional hypothesis based on existing knowledge.

Develop a Theoretical Framework:

Build a theoretical framework that outlines the expected causal relationship between variables. The theoretical framework should provide a clear rationale for predicting the direction of the relationship based on established theories or concepts.

Conduct Pilot Studies:

Conducting pilot studies or preliminary research can provide valuable insights and data to inform the formulation of a directional hypothesis. Initial findings can help researchers identify patterns or relationships that support a specific predicted direction.

Seek Expert Input:

Seek input from experts or colleagues in the field who have expertise in the area of study. Discuss the research question and hypothesis with them to obtain valuable insights, perspectives, and feedback that can help refine and improve the directional hypothesis.

Clearly Define Variables:

Clearly define and operationalize the variables in the hypothesis to ensure precision and clarity. This will help avoid ambiguity and ensure that the hypothesis is testable and measurable.

Strategies to Improve Non-Directional Hypotheses:

Preliminary exploration:.

Conduct initial exploratory research to gather preliminary data and insights on the relationship or difference between variables. This can provide a foundation for formulating a non-directional hypothesis based on observed patterns or trends.

Analyze Existing Data:

Analyze existing datasets to identify potential relationships or differences. Exploratory data analysis techniques such as data visualization, descriptive statistics, and correlation analysis can help uncover initial insights that can guide the formulation of a non-directional hypothesis.

Use Exploratory Research Designs:

Employ exploratory research designs such as qualitative studies, case studies, or grounded theory approaches. These designs allow researchers to gather rich data and explore relationships or differences without preconceived notions about the direction.

Consider Alternative Explanations:

When formulating a non-directional hypothesis, consider alternative explanations or potential factors that may influence the relationship or difference between variables. This can help ensure a comprehensive and nuanced understanding of the phenomenon under investigation.

Refine Based on Initial Findings:

Refine the non-directional hypothesis based on initial findings and observations from exploratory analyses. These findings can guide the formulation of more specific hypotheses in subsequent studies or inform the direction of further research.

In conclusion, both directional and non-directional hypotheses have their merits and are valuable in different research contexts.

 Here’s a summary of the key points regarding directional and non-directional hypotheses:

  • A directional hypothesis makes a specific prediction about the direction of the relationship or difference between variables.
  • It is appropriate when there is a clear theoretical or empirical basis for predicting the direction.
  • Directional hypotheses provide a focused approach, guiding the research towards confirming or refuting a specific predicted direction.
  • They are useful in studies where cause-and-effect relationships are being examined or when specific practical implications are desired.
  • Directional hypotheses require careful consideration of prior research, theoretical frameworks, and available evidence.
  • A non-directional hypothesis does not specify the direction of the relationship or difference between variables.
  • It is employed when there is limited prior knowledge, conflicting findings, or a desire for exploratory analysis.
  • Non-directional hypotheses allow for flexibility and open-mindedness in exploring the data, considering both positive and negative associations or differences.
  • They are suitable for preliminary studies, exploratory research, or when the research question does not have a clear predicted direction.
  • Non-directional hypotheses are beneficial for generating new hypotheses, replication studies, and enhancing generalizability.

In both cases, it is essential to ensure that hypotheses are clear, testable, and aligned with the research objectives. Researchers should also be open to revising and refining hypotheses based on the findings and feedback obtained during the research process. The choice between a directional and non-directional hypothesis depends on factors such as the research question, available literature, theoretical frameworks, and the specific objectives of the study. Researchers should carefully consider these factors to determine the most appropriate type of hypothesis to use in their research

psychologyrocks

Hypotheses; directional and non-directional, what is the difference between an experimental and an alternative hypothesis.

Nothing much! If the study is a true experiment then we can call the hypothesis “an experimental hypothesis”, a prediction is made about how the IV causes an effect on the DV. In a study which does not involve the direct manipulation of an IV, i.e. a natural or quasi-experiment or any other quantitative research method (e.g. survey) has been used, then we call it an “alternative hypothesis”, it is the alternative to the null.

Directional hypothesis: A directional (or one-tailed hypothesis) states which way you think the results are going to go, for example in an experimental study we might say…”Participants who have been deprived of sleep for 24 hours will have more cold symptoms the week after exposure to a virus than participants who have not been sleep deprived”; the hypothesis compares the two groups/conditions and states which one will ….have more/less, be quicker/slower, etc.

If we had a correlational study, the directional hypothesis would state whether we expect a positive or a negative correlation, we are stating how the two variables will be related to each other, e.g. there will be a positive correlation between the number of stressful life events experienced in the last year and the number of coughs and colds suffered, whereby the more life events you have suffered the more coughs and cold you will have had”. The directional hypothesis can also state a negative correlation, e.g. the higher the number of face-book friends, the lower the life satisfaction score “

Non-directional hypothesis: A non-directional (or two tailed hypothesis) simply states that there will be a difference between the two groups/conditions but does not say which will be greater/smaller, quicker/slower etc. Using our example above we would say “There will be a difference between the number of cold symptoms experienced in the following week after exposure to a virus for those participants who have been sleep deprived for 24 hours compared with those who have not been sleep deprived for 24 hours.”

When the study is correlational, we simply state that variables will be correlated but do not state whether the relationship will be positive or negative, e.g. there will be a significant correlation between variable A and variable B.

Null hypothesis The null hypothesis states that the alternative or experimental hypothesis is NOT the case, if your experimental hypothesis was directional you would say…

Participants who have been deprived of sleep for 24 hours will NOT have more cold symptoms in the following week after exposure to a virus than participants who have not been sleep deprived and any difference that does arise will be due to chance alone.

or with a directional correlational hypothesis….

There will NOT be a positive correlation between the number of stress life events experienced in the last year and the number of coughs and colds suffered, whereby the more life events you have suffered the more coughs and cold you will have had”

With a non-directional or  two tailed hypothesis…

There will be NO difference between the number of cold symptoms experienced in the following week after exposure to a virus for those participants who have been sleep deprived for 24 hours compared with those who have not been sleep deprived for 24 hours.

or for a correlational …

there will be NO correlation between variable A and variable B.

When it comes to conducting an inferential stats test, if you have a directional hypothesis , you must do a one tailed test to find out whether your observed value is significant. If you have a non-directional hypothesis , you must do a two tailed test .

Exam Techniques/Advice

  • Remember, a decent hypothesis will contain two variables, in the case of an experimental hypothesis there will be an IV and a DV; in a correlational hypothesis there will be two co-variables
  • both variables need to be fully operationalised to score the marks, that is you need to be very clear and specific about what you mean by your IV and your DV; if someone wanted to repeat your study, they should be able to look at your hypothesis and know exactly what to change between the two groups/conditions and exactly what to measure (including any units/explanation of rating scales etc, e.g. “where 1 is low and 7 is high”)
  • double check the question, did it ask for a directional or non-directional hypothesis?
  • if you were asked for a null hypothesis, make sure you always include the phrase “and any difference/correlation (is your study experimental or correlational?) that does arise will be due to chance alone”

Practice Questions:

  • Mr Faraz wants to compare the levels of attendance between his psychology group and those of Mr Simon, who teaches a different psychology group. Which of the following is a suitable directional (one tailed) hypothesis for Mr Faraz’s investigation?

A There will be a difference in the levels of attendance between the two psychology groups.

B Students’ level of attendance will be higher in Mr Faraz’s group than Mr Simon’s group.

C Any difference in the levels of attendance between the two psychology groups is due to chance.

D The level of attendance of the students will depend upon who is teaching the groups.

2. Tracy works for the local council. The council is thinking about reducing the number of people it employs to pick up litter from the street. Tracy has been asked to carry out a study to see if having the streets cleaned at less regular intervals will affect the amount of litter the public will drop. She studies a street to compare how much litter is dropped at two different times, once when it has just been cleaned and once after it has not been cleaned for a month.

Write a fully operationalised non-directional (two-tailed) hypothesis for Tracy’s study. (2)

3. Jamila is conducting a practical investigation to look at gender differences in carrying out visuo-spatial tasks. She decides to give males and females a jigsaw puzzle and will time them to see who completes it the fastest. She uses a random sample of pupils from a local school to get her participants.

(a) Write a fully operationalised directional (one tailed) hypothesis for Jamila’s study. (2) (b) Outline one strength and one weakness of the random sampling method. You may refer to Jamila’s use of this type of sampling in your answer. (4)

4. Which of the following is a non-directional (two tailed) hypothesis?

A There is a difference in driving ability with men being better drivers than women

B Women are better at concentrating on more than one thing at a time than men

C Women spend more time doing the cooking and cleaning than men

D There is a difference in the number of men and women who participate in sports

Revision Activities

writing-hypotheses-revision-sheet

Quizizz link for teachers: https://quizizz.com/admin/quiz/5bf03f51add785001bc5a09e

By Psychstix by Mandy wood

Share this:

' src=

  • Already have a WordPress.com account? Log in now.
  • Subscribe Subscribed
  • Copy shortlink
  • Report this content
  • View post in Reader
  • Manage subscriptions
  • Collapse this bar

Directional and non-directional hypothesis: A Comprehensive Guide

Karolina Konopka

Customer support manager

Karolina Konopka

Want to talk with us?

In the world of research and statistical analysis, hypotheses play a crucial role in formulating and testing scientific claims. Understanding the differences between directional and non-directional hypothesis is essential for designing sound experiments and drawing accurate conclusions. Whether you’re a student, researcher, or simply curious about the foundations of hypothesis testing, this guide will equip you with the knowledge and tools to navigate this fundamental aspect of scientific inquiry.

Understanding Directional Hypothesis

Understanding directional hypotheses is crucial for conducting hypothesis-driven research, as they guide the selection of appropriate statistical tests and aid in the interpretation of results. By incorporating directional hypotheses, researchers can make more precise predictions, contribute to scientific knowledge, and advance their fields of study.

Definition of directional hypothesis

Directional hypotheses, also known as one-tailed hypotheses, are statements in research that make specific predictions about the direction of a relationship or difference between variables. Unlike non-directional hypotheses, which simply state that there is a relationship or difference without specifying its direction, directional hypotheses provide a focused and precise expectation.

A directional hypothesis predicts either a positive or negative relationship between variables or predicts that one group will perform better than another. It asserts a specific direction of effect or outcome. For example, a directional hypothesis could state that “increased exposure to sunlight will lead to an improvement in mood” or “participants who receive the experimental treatment will exhibit higher levels of cognitive performance compared to the control group.”

Directional hypotheses are formulated based on existing theory, prior research, or logical reasoning, and they guide the researcher’s expectations and analysis. They allow for more targeted predictions and enable researchers to test specific hypotheses using appropriate statistical tests.

The role of directional hypothesis in research

Directional hypotheses also play a significant role in research surveys. Let’s explore their role specifically in the context of survey research:

  • Objective-driven surveys : Directional hypotheses help align survey research with specific objectives. By formulating directional hypotheses, researchers can focus on gathering data that directly addresses the predicted relationship or difference between variables of interest.
  • Question design and measurement : Directional hypotheses guide the design of survey question types and the selection of appropriate measurement scales. They ensure that the questions are tailored to capture the specific aspects related to the predicted direction, enabling researchers to obtain more targeted and relevant data from survey respondents.
  • Data analysis and interpretation : Directional hypotheses assist in data analysis by directing researchers towards appropriate statistical tests and methods. Researchers can analyze the survey data to specifically test the predicted relationship or difference, enhancing the accuracy and reliability of their findings. The results can then be interpreted within the context of the directional hypothesis, providing more meaningful insights.
  • Practical implications and decision-making : Directional hypotheses in surveys often have practical implications. When the predicted relationship or difference is confirmed, it informs decision-making processes, program development, or interventions. The survey findings based on directional hypotheses can guide organizations, policymakers, or practitioners in making informed choices to achieve desired outcomes.
  • Replication and further research : Directional hypotheses in survey research contribute to the replication and extension of studies. Researchers can replicate the survey with different populations or contexts to assess the generalizability of the predicted relationships. Furthermore, if the directional hypothesis is supported, it encourages further research to explore underlying mechanisms or boundary conditions.

By incorporating directional hypotheses in survey research, researchers can align their objectives, design effective surveys, conduct focused data analysis, and derive practical insights. They provide a framework for organizing survey research and contribute to the accumulation of knowledge in the field.

Examples of research questions for directional hypothesis

Here are some examples of research questions that lend themselves to directional hypotheses:

  • Does increased daily exercise lead to a decrease in body weight among sedentary adults?
  • Is there a positive relationship between study hours and academic performance among college students?
  • Does exposure to violent video games result in an increase in aggressive behavior among adolescents?
  • Does the implementation of a mindfulness-based intervention lead to a reduction in stress levels among working professionals?
  • Is there a difference in customer satisfaction between Product A and Product B, with Product A expected to have higher satisfaction ratings?
  • Does the use of social media influence self-esteem levels, with higher social media usage associated with lower self-esteem?
  • Is there a negative relationship between job satisfaction and employee turnover, indicating that lower job satisfaction leads to higher turnover rates?
  • Does the administration of a specific medication result in a decrease in symptoms among individuals with a particular medical condition?
  • Does increased access to early childhood education lead to improved cognitive development in preschool-aged children?
  • Is there a difference in purchase intention between advertisements with celebrity endorsements and advertisements without, with celebrity endorsements expected to have a higher impact?

These research questions generate specific predictions about the direction of the relationship or difference between variables and can be tested using appropriate research methods and statistical analyses.

Definition of non-directional hypothesis

Non-directional hypotheses, also known as two-tailed hypotheses, are statements in research that indicate the presence of a relationship or difference between variables without specifying the direction of the effect. Instead of making predictions about the specific direction of the relationship or difference, non-directional hypotheses simply state that there is an association or distinction between the variables of interest.

Non-directional hypotheses are often used when there is no prior theoretical basis or clear expectation about the direction of the relationship. They leave the possibility open for either a positive or negative relationship, or for both groups to differ in some way without specifying which group will perform better or worse.

Advantages and utility of non-directional hypothesis

Non-directional hypotheses in survey s offer several advantages and utilities, providing flexibility and comprehensive analysis of survey data. Here are some of the key advantages and utilities of using non-directional hypotheses in surveys:

  • Exploration of Relationships : Non-directional hypotheses allow researchers to explore and examine relationships between variables without assuming a specific direction. This is particularly useful in surveys where the relationship between variables may not be well-known or there may be conflicting evidence regarding the direction of the effect.
  • Flexibility in Question Design : With non-directional hypotheses, survey questions can be designed to measure the relationship between variables without being biased towards a particular outcome. This flexibility allows researchers to collect data and analyze the results more objectively.
  • Open to Unexpected Findings : Non-directional hypotheses enable researchers to be open to unexpected or surprising findings in survey data. By not committing to a specific direction of the effect, researchers can identify and explore relationships that may not have been initially anticipated, leading to new insights and discoveries.
  • Comprehensive Analysis : Non-directional hypotheses promote comprehensive analysis of survey data by considering the possibility of an effect in either direction. Researchers can assess the magnitude and significance of relationships without limiting their analysis to only one possible outcome.
  • S tatistical Validity : Non-directional hypotheses in surveys allow for the use of two-tailed statistical tests, which provide a more conservative and robust assessment of significance. Two-tailed tests consider both positive and negative deviations from the null hypothesis, ensuring accurate and reliable statistical analysis of survey data.
  • Exploratory Research : Non-directional hypotheses are particularly useful in exploratory research, where the goal is to gather initial insights and generate hypotheses. Surveys with non-directional hypotheses can help researchers explore various relationships and identify patterns that can guide further research or hypothesis development.

It is worth noting that the choice between directional and non-directional hypotheses in surveys depends on the research objectives, existing knowledge, and the specific variables being investigated. Researchers should carefully consider the advantages and limitations of each approach and select the one that aligns best with their research goals and survey design.

  • Share with others
  • Twitter Twitter Icon
  • LinkedIn LinkedIn Icon

Related posts

Mastering matrix questions: the complete guide + one special technique, microsurveys: complete guide, recurring surveys: the ultimate guide, close-ended questions: definition, types, and examples, top 10 typeform alternatives and competitors, 15 best google forms alternatives and competitors, get answers today.

  • No credit card required
  • No time limit on Free plan

You can modify this template in every possible way.

All templates work great on every device.

  • Skip to secondary menu
  • Skip to main content
  • Skip to primary sidebar

Statistics By Jim

Making statistics intuitive

One-Tailed and Two-Tailed Hypothesis Tests Explained

By Jim Frost 61 Comments

Choosing whether to perform a one-tailed or a two-tailed hypothesis test is one of the methodology decisions you might need to make for your statistical analysis. This choice can have critical implications for the types of effects it can detect, the statistical power of the test, and potential errors.

In this post, you’ll learn about the differences between one-tailed and two-tailed hypothesis tests and their advantages and disadvantages. I include examples of both types of statistical tests. In my next post, I cover the decision between one and two-tailed tests in more detail.

What Are Tails in a Hypothesis Test?

First, we need to cover some background material to understand the tails in a test. Typically, hypothesis tests take all of the sample data and convert it to a single value, which is known as a test statistic. You’re probably already familiar with some test statistics. For example, t-tests calculate t-values . F-tests, such as ANOVA, generate F-values . The chi-square test of independence and some distribution tests produce chi-square values. All of these values are test statistics. For more information, read my post about Test Statistics .

These test statistics follow a sampling distribution. Probability distribution plots display the probabilities of obtaining test statistic values when the null hypothesis is correct. On a probability distribution plot, the portion of the shaded area under the curve represents the probability that a value will fall within that range.

The graph below displays a sampling distribution for t-values. The two shaded regions cover the two-tails of the distribution.

Plot that display critical regions in the two tails of the distribution.

Keep in mind that this t-distribution assumes that the null hypothesis is correct for the population. Consequently, the peak (most likely value) of the distribution occurs at t=0, which represents the null hypothesis in a t-test. Typically, the null hypothesis states that there is no effect. As t-values move further away from zero, it represents larger effect sizes. When the null hypothesis is true for the population, obtaining samples that exhibit a large apparent effect becomes less likely, which is why the probabilities taper off for t-values further from zero.

Related posts : How t-Tests Work and Understanding Probability Distributions

Critical Regions in a Hypothesis Test

In hypothesis tests, critical regions are ranges of the distributions where the values represent statistically significant results. Analysts define the size and location of the critical regions by specifying both the significance level (alpha) and whether the test is one-tailed or two-tailed.

Consider the following two facts:

  • The significance level is the probability of rejecting a null hypothesis that is correct.
  • The sampling distribution for a test statistic assumes that the null hypothesis is correct.

Consequently, to represent the critical regions on the distribution for a test statistic, you merely shade the appropriate percentage of the distribution. For the common significance level of 0.05, you shade 5% of the distribution.

Related posts : Significance Levels and P-values and T-Distribution Table of Critical Values

Two-Tailed Hypothesis Tests

Two-tailed hypothesis tests are also known as nondirectional and two-sided tests because you can test for effects in both directions. When you perform a two-tailed test, you split the significance level percentage between both tails of the distribution. In the example below, I use an alpha of 5% and the distribution has two shaded regions of 2.5% (2 * 2.5% = 5%).

When a test statistic falls in either critical region, your sample data are sufficiently incompatible with the null hypothesis that you can reject it for the population.

In a two-tailed test, the generic null and alternative hypotheses are the following:

  • Null : The effect equals zero.
  • Alternative :  The effect does not equal zero.

The specifics of the hypotheses depend on the type of test you perform because you might be assessing means, proportions, or rates.

Example of a two-tailed 1-sample t-test

Suppose we perform a two-sided 1-sample t-test where we compare the mean strength (4.1) of parts from a supplier to a target value (5). We use a two-tailed test because we care whether the mean is greater than or less than the target value.

To interpret the results, simply compare the p-value to your significance level. If the p-value is less than the significance level, you know that the test statistic fell into one of the critical regions, but which one? Just look at the estimated effect. In the output below, the t-value is negative, so we know that the test statistic fell in the critical region in the left tail of the distribution, indicating the mean is less than the target value. Now we know this difference is statistically significant.

Statistical output from a two-tailed 1-sample t-test.

We can conclude that the population mean for part strength is less than the target value. However, the test had the capacity to detect a positive difference as well. You can also assess the confidence interval. With a two-tailed hypothesis test, you’ll obtain a two-sided confidence interval. The confidence interval tells us that the population mean is likely to fall between 3.372 and 4.828. This range excludes the target value (5), which is another indicator of significance.

Advantages of two-tailed hypothesis tests

You can detect both positive and negative effects. Two-tailed tests are standard in scientific research where discovering any type of effect is usually of interest to researchers.

One-Tailed Hypothesis Tests

One-tailed hypothesis tests are also known as directional and one-sided tests because you can test for effects in only one direction. When you perform a one-tailed test, the entire significance level percentage goes into the extreme end of one tail of the distribution.

In the examples below, I use an alpha of 5%. Each distribution has one shaded region of 5%. When you perform a one-tailed test, you must determine whether the critical region is in the left tail or the right tail. The test can detect an effect only in the direction that has the critical region. It has absolutely no capacity to detect an effect in the other direction.

In a one-tailed test, you have two options for the null and alternative hypotheses, which corresponds to where you place the critical region.

You can choose either of the following sets of generic hypotheses:

  • Null : The effect is less than or equal to zero.
  • Alternative : The effect is greater than zero.

Plot that displays a single critical region for a one-tailed test.

  • Null : The effect is greater than or equal to zero.
  • Alternative : The effect is less than zero.

Plot that displays a single critical region in the left tail for a one-tailed test.

Again, the specifics of the hypotheses depend on the type of test you perform.

Notice how for both possible null hypotheses the tests can’t distinguish between zero and an effect in a particular direction. For example, in the example directly above, the null combines “the effect is greater than or equal to zero” into a single category. That test can’t differentiate between zero and greater than zero.

Example of a one-tailed 1-sample t-test

Suppose we perform a one-tailed 1-sample t-test. We’ll use a similar scenario as before where we compare the mean strength of parts from a supplier (102) to a target value (100). Imagine that we are considering a new parts supplier. We will use them only if the mean strength of their parts is greater than our target value. There is no need for us to differentiate between whether their parts are equally strong or less strong than the target value—either way we’d just stick with our current supplier.

Consequently, we’ll choose the alternative hypothesis that states the mean difference is greater than zero (Population mean – Target value > 0). The null hypothesis states that the difference between the population mean and target value is less than or equal to zero.

Statistical output for a one-tailed 1-sample t-test.

To interpret the results, compare the p-value to your significance level. If the p-value is less than the significance level, you know that the test statistic fell into the critical region. For this study, the statistically significant result supports the notion that the population mean is greater than the target value of 100.

Confidence intervals for a one-tailed test are similarly one-sided. You’ll obtain either an upper bound or a lower bound. In this case, we get a lower bound, which indicates that the population mean is likely to be greater than or equal to 100.631. There is no upper limit to this range.

A lower-bound matches our goal of determining whether the new parts are stronger than our target value. The fact that the lower bound (100.631) is higher than the target value (100) indicates that these results are statistically significant.

This test is unable to detect a negative difference even when the sample mean represents a very negative effect.

Advantages and disadvantages of one-tailed hypothesis tests

One-tailed tests have more statistical power to detect an effect in one direction than a two-tailed test with the same design and significance level. One-tailed tests occur most frequently for studies where one of the following is true:

  • Effects can exist in only one direction.
  • Effects can exist in both directions but the researchers only care about an effect in one direction. There is no drawback to failing to detect an effect in the other direction. (Not recommended.)

The disadvantage of one-tailed tests is that they have no statistical power to detect an effect in the other direction.

As part of your pre-study planning process, determine whether you’ll use the one- or two-tailed version of a hypothesis test. To learn more about this planning process, read 5 Steps for Conducting Scientific Studies with Statistical Analyses .

This post explains the differences between one-tailed and two-tailed statistical hypothesis tests. How these forms of hypothesis tests function is clear and based on mathematics. However, there is some debate about when you can use one-tailed tests. My next post explores this decision in much more depth and explains the different schools of thought and my opinion on the matter— When Can I Use One-Tailed Hypothesis Tests .

If you’re learning about hypothesis testing and like the approach I use in my blog, check out my Hypothesis Testing book! You can find it at Amazon and other retailers.

Cover image of my Hypothesis Testing: An Intuitive Guide ebook.

Share this:

non directional two tailed hypothesis

Reader Interactions

' src=

August 23, 2024 at 1:28 pm

Thank so much. This is very helpfull

' src=

June 26, 2022 at 12:14 pm

Hi, Can help me with figuring out the null and alternative hypothesis of the following statement? Some claimed that the real average expenditure on beverage by general people is at least $10.

' src=

February 19, 2022 at 6:02 am

thank you for the thoroughly explanation, I’m still strugling to wrap my mind around the t-table and the relation between the alpha values for one or two tail probability and the confidence levels on the bottom (I’m understanding it so wrongly that for me it should be the oposite, like one tail 0,05 should correspond 95% CI and two tailed 0,025 should correspond to 95% because then you got the 2,5% on each side). In my mind if I picture the one tail diagram with an alpha of 0,05 I see the rest 95% inside the diagram, but for a one tail I only see 90% CI paired with a 5% alpha… where did the other 5% go? I tried to understand when you said we should just double the alpha for a one tail probability in order to find the CI but I still cant picture it. I have been trying to understand this. Like if you only have one tail and there is 0,05, shouldn’t the rest be on the other side? why is it then 90%… I know I’m missing a point and I can’t figure it out and it’s so frustrating…

' src=

February 23, 2022 at 10:01 pm

The alpha is the total shaded area. So, if the alpha = 0.05, you know that 5% of the distribution is shaded. The number of tails tells you how to divide the shaded areas. Is it all in one region (1-tailed) or do you split the shaded regions in two (2-tailed)?

So, for a one-tailed test with an alpha of 0.05, the 5% shading is all in one tail. If alpha = 0.10, then it’s 10% on one side. If it’s two-tailed, then you need to split that 10% into two–5% in both tails. Hence, the 5% in a one-tailed test is the same as a two-tailed test with an alpha of 0.10 because that test has the same 5% on one side (but there’s another 5% in the other tail).

It’s similar for CIs. However, for CIs, you shade the middle rather than the extremities. I write about that in one my articles about hypothesis testing and confidence intervals .

I’m not sure if I’m answering your question or not.

' src=

February 17, 2022 at 1:46 pm

I ran a post hoc Dunnett’s test alpha=0.05 after a significant Anova test in Proc Mixed using SAS. I want to determine if the means for treatment (t1, t2, t3) is significantly less than the means for control (p=pathogen). The code for the dunnett’s test is – LSmeans trt / diff=controll (‘P’) adjust=dunnett CL plot=control; I think the lower bound one tailed test is the correct test to run but I’m not 100% sure. I’m finding conflicting information online. In the output table for the dunnett’s test the mean difference between the control and the treatments is t1=9.8, t2=64.2, and t3=56.5. The control mean estimate is 90.5. The adjusted p-value by treatment is t1(p=0.5734), t2 (p=.0154) and t3(p=.0245). The adjusted lower bound confidence limit in order from t1-t3 is -38.8, 13.4, and 7.9. The adjusted upper bound for all test is infinity. The graphical output for the dunnett’s test in SAS is difficult to understand for those of us who are beginner SAS users. All treatments appear as a vertical line below the the horizontal line for control at 90.5 with t2 and t3 in the shaded area. For treatment 1 the shaded area is above the line for control. Looking at just the output table I would say that t2 and t3 are significantly lower than the control. I guess I would like to know if my interpretation of the outputs is correct that treatments 2 and 3 are statistically significantly lower than the control? Should I have used an upper bound one tailed test instead?

' src=

November 10, 2021 at 1:00 am

Thanks Jim. Please help me understand how a two tailed testing can be used to minimize errors in research

' src=

July 1, 2021 at 9:19 am

Hi Jim, Thanks for posting such a thorough and well-written explanation. It was extremely useful to clear up some doubts.

' src=

May 7, 2021 at 4:27 pm

Hi Jim, I followed your instructions for the Excel add-in. Thank you. I am very new to statistics and sort of enjoy it as I enter week number two in my class. I am to select if three scenarios call for a one or two-tailed test is required and why. The problem is stated:

30% of mole biopsies are unnecessary. Last month at his clinic, 210 out of 634 had benign biopsy results. Is there enough evidence to reject the dermatologist’s claim?

Part two, the wording changes to “more than of 30% of biopsies,” and part three, the wording changes to “less than 30% of biopsies…”

I am not asking for the problem to be solved for me, but I cannot seem to find direction needed. I know the elements i am dealing with are =30%, greater than 30%, and less than 30%. 210 and 634. I just don’t know what to with the information. I can’t seem to find an example of a similar problem to work with.

May 9, 2021 at 9:22 pm

As I detail in this post, a two-tailed test tells you whether an effect exists in either direction. Or, is it different from the null value in either direction. For the first example, the wording suggests you’d need a two-tailed test to determine whether the population proportion is ≠ 30%. Whenever you just need to know ≠, it suggests a two-tailed test because you’re covering both directions.

For part two, because it’s in one direction (greater than), you need a one-tailed test. Same for part three but it’s less than. Look in this blog post to see how you’d construct the null and alternative hypotheses for these cases. Note that you’re working with a proportion rather than the mean, but the principles are the same! Just plug your scenario and the concept of proportion into the wording I use for the hypotheses.

I hope that helps!

' src=

April 11, 2021 at 9:30 am

Hello Jim, great website! I am using a statistics program (SPSS) that does NOT compute one-tailed t-tests. I am trying to compare two independent groups and have justifiable reasons why I only care about one direction. Can I do the following? Use SPSS for two-tailed tests to calculate the t & p values. Then report the p-value as p/2 when it is in the predicted direction (e.g , SPSS says p = .04, so I report p = .02), and report the p-value as 1 – (p/2) when it is in the opposite direction (e.g., SPSS says p = .04, so I report p = .98)? If that is incorrect, what do you suggest (hopefully besides changing statistics programs)? Also, if I want to report confidence intervals, I realize that I would only have an upper or lower bound, but can I use the CI’s from SPSS to compute that? Thank you very much!

April 11, 2021 at 5:42 pm

Yes, for p-values, that’s absolutely correct for both cases.

For confidence intervals, if you take one endpoint of a two-side CI, it becomes a one-side bound with half the confidence level.

Consequently, to obtain a one-sided bound with your desired confidence level, you need to take your desired significance level (e.g., 0.05) and double it. Then subtract it from 1. So, if you’re using a significance level of 0.05, double that to 0.10 and then subtract from 1 (1 – 0.10 = 0.90). 90% is the confidence level you want to use for a two-sided test. After obtaining the two-sided CI, use one of the endpoints depending on the direction of your hypothesis (i.e., upper or lower bound). That’s produces the one-sided the bound with the confidence level that you want. For our example, we calculated a 95% one-sided bound.

' src=

March 3, 2021 at 8:27 am

Hi Jim. I used the one-tailed(right) statistical test to determine an anomaly in the below problem statement: On a daily basis, I calculate the (mapped_%) in a common field between two tables.

The way I used the t-test is: On any particular day, I calculate the sample_mean, S.D and sample_count (n=30) for the last 30 days including the current day. My null hypothesis, H0 (pop. mean)=95 and H1>95 (alternate hypothesis). So, I calculate the t-stat based on the sample_mean, pop.mean, sample S.D and n. I then choose the t-crit value for 0.05 from my t-ditribution table for dof(n-1). On the current day if my abs.(t-stat)>t-crit, then I reject the null hypothesis and I say the mapped_pct on that day has passed the t-test.

I get some weird results here, where if my mapped_pct is as low as 6%-8% in all the past 30 days, the t-test still gets a “pass” result. Could you help on this? If my hypothesis needs to be changed.

I would basically look for the mapped_pct >95, if it worked on a static trigger. How can I use the t-test effectively in this problem statement?

' src=

December 18, 2020 at 8:23 pm

Hello Dr. Jim, I am wondering if there is evidence in one of your books or other source you could provide, which supports that it is OK not to divide alpha level by 2 in one-tailed hypotheses. I need the source for supporting evidence in a Portfolio exercise and couldn’t find one.

I am grateful for your reply and for your statistics knowledge sharing!

' src=

November 27, 2020 at 10:31 pm

If I did a one directional F test ANOVA(one tail ) and wanted to calculate a confidence interval for each individual groups (3) mean . Would I use a one tailed or two tailed t , within my confidence interval .

November 29, 2020 at 2:36 am

Hi Bashiru,

F-tests for ANOVA will always be one-tailed for the reasons I discuss in this post. To learn more about, read my post about F-tests in ANOVA .

For the differences between my groups, I would not use t-tests because the family-wise error rate quickly grows out of hand. To learn more about how to compare group means while controlling the familywise error rate, read my post about using post hoc tests with ANOVA . Typically, these are two-side intervals but you’d be able to use one-sided.

' src=

November 26, 2020 at 10:51 am

Hi Jim, I had a question about the formulation of the hypotheses. When you want to test if a beta = 1 or a beta = 0. What will be the null hypotheses? I’m having trouble with finding out. Because in most cases beta = 0 is the null hypotheses but in this case you want to test if beta = 0. so i’m having my doubts can it in this case be the alternative hypotheses or is it still the null hypotheses?

Kind regards, Noa

November 27, 2020 at 1:21 am

Typically, the null hypothesis represents no effect or no relationship. As an analyst, you’re hoping that your data have enough evidence to reject the null and favor the alternative.

Assuming you’re referring to beta as in regression coefficients, zero represents no relationship. Consequently, beta = 0 is the null hypothesis.

You might hope that beta = 1, but you don’t usually include that in your alternative hypotheses. The alternative hypothesis usually states that it does not equal no effect. In other words, there is an effect but it doesn’t state what it is.

There are some exceptions to the above but I’m writing about the standard case.

' src=

November 22, 2020 at 8:46 am

Your articles are a help to intro to econometrics students. Keep up the good work! More power to you!

' src=

November 6, 2020 at 11:25 pm

Hello Jim. Can you help me with these please?

Write the null and alternative hypothesis using a 1-tailed and 2-tailed test for each problem. (In paragraph and symbols)

A teacher wants to know if there is a significant difference in the performance in MAT C313 between her morning and afternoon classes.

It is known that in our university canteen, the average waiting time for a customer to receive and pay for his/her order is 20 minutes. Additional personnel has been added and now the management wants to know if the average waiting time had been reduced.

November 8, 2020 at 12:29 am

I cover how to write the hypotheses for the different types of tests in this post. So, you just need to figure which type of test you need to use. In your case, you want to determine whether the mean waiting time is less than the target value of 20 minutes. That’s a 1-sample t-test because you’re comparing a mean to a target value (20 minutes). You specifically want to determine whether the mean is less than the target value. So, that’s a one-tailed test. And, you’re looking for a mean that is “less than” the target.

So, go to the one-tailed section in the post and look for the hypotheses for the effect being less than. That’s the one with the critical region on the left side of the curve.

Now, you need include your own information. In your case, you’re comparing the sample estimate to a population mean of 20. The 20 minutes is your null hypothesis value. Use the symbol mu μ to represent the population mean.

You put all that together and you get the following:

Null: μ ≥ 20 Alternative: μ 0 to denote the null hypothesis and H 1 or H A to denote the alternative hypothesis if that’s what you been using in class.

' src=

October 17, 2020 at 12:11 pm

I was just wondering if you could please help with clarifying what the hypothesises would be for say income for gamblers and, age of gamblers. I am struggling to find which means would be compared.

October 17, 2020 at 7:05 pm

Those are both continuous variables, so you’d use either correlation or regression for them. For both of those analyses, the hypotheses are the following:

Null : The correlation or regression coefficient equals zero (i.e., there is no relationship between the variables) Alternative : The coefficient does not equal zero (i.e., there is a relationship between the variables.)

When the p-value is less than your significance level, you reject the null and conclude that a relationship exists.

' src=

October 17, 2020 at 3:05 am

I was ask to choose and justify the reason between a one tailed and two tailed test for dummy variables, how do I do that and what does it mean?

October 17, 2020 at 7:11 pm

I don’t have enough information to answer your question. A dummy variable is also known as an indicator variable, which is a binary variable that indicates the presence or absence of a condition or characteristic. If you’re using this variable in a hypothesis test, I’d presume that you’re using a proportions test, which is based on the binomial distribution for binary data.

Choosing between a one-tailed or two-tailed test depends on subject area issues and, possibly, your research objectives. Typically, use a two-tailed test unless you have a very good reason to use a one-tailed test. To understand when you might use a one-tailed test, read my post about when to use a one-tailed hypothesis test .

' src=

October 16, 2020 at 2:07 pm

In your one-tailed example, Minitab describes the hypotheses as “Test of mu = 100 vs > 100”. Any idea why Minitab says the null is “=” rather than “= or less than”? No ASCII character for it?

October 16, 2020 at 4:20 pm

I’m not entirely sure even though I used to work there! I know we had some discussions about how to represent that hypothesis but I don’t recall the exact reasoning. I suspect that it has to do with the conclusions that you can draw. Let’s focus on the failing to reject the null hypothesis. If the test statistic falls in that region (i.e., it is not significant), you fail to reject the null. In this case, all you know is that you have insufficient evidence to say it is different than 100. I’m pretty sure that’s why they use the equal sign because it might as well be one.

Mathematically, I think using ≤ is more accurate, which you can really see when you look at the distribution plots. That’s why I phrase the hypotheses using ≤ or ≥ as needed. However, in terms of the interpretation, the “less than” portion doesn’t really add anything of importance. You can conclude that its equal to 100 or greater than 100, but not less than 100.

' src=

October 15, 2020 at 5:46 am

Thank you so much for your timely feedback. It helps a lot

October 14, 2020 at 10:47 am

How can i use one tailed test at 5% alpha on this problem?

A manufacturer of cellular phone batteries claims that when fully charged, the mean life of his product lasts for 26 hours with a standard deviation of 5 hours. Mr X, a regular distributor, randomly picked and tested 35 of the batteries. His test showed that the average life of his sample is 25.5 hours. Is there a significant difference between the average life of all the manufacturer’s batteries and the average battery life of his sample?

October 14, 2020 at 8:22 pm

I don’t think you’d want to use a one-tailed test. The goal is to determine whether the sample is significantly different than the manufacturer’s population average. You’re not saying significantly greater than or less than, which would be a one-tailed test. As phrased, you want a two-tailed test because it can detect a difference in either direct.

It sounds like you need to use a 1-sample t-test to test the mean. During this test, enter 26 as the test mean. The procedure will tell you if the sample mean of 25.5 hours is a significantly different from that test mean. Similarly, you’d need a one variance test to determine whether the sample standard deviation is significantly different from the test value of 5 hours.

For both of these tests, compare the p-value to your alpha of 0.05. If the p-value is less than this value, your results are statistically significant.

' src=

September 22, 2020 at 4:16 am

Hi Jim, I didn’t get an idea that when to use two tail test and one tail test. Will you please explain?

September 22, 2020 at 10:05 pm

I have a complete article dedicated to that: When Can I Use One-Tailed Tests .

Basically, start with the assumption that you’ll use a two-tailed test but then consider scenarios where a one-tailed test can be appropriate. I talk about all of that in the article.

If you have questions after reading that, please don’t hesitate to ask!

' src=

July 31, 2020 at 12:33 pm

Thank you so so much for this webpage.

I have two scenarios that I need some clarification. I will really appreciate it if you can take a look:

So I have several of materials that I know when they are tested after production. My hypothesis is that the earlier they are tested after production, the higher the mean value I should expect. At the same time, the later they are tested after production, the lower the mean value. Since this is more like a “greater or lesser” situation, I should use one tail. Is that the correct approach?

On the other hand, I have several mix of materials that I don’t know when they are tested after production. I only know the mean values of the test. And I only want to know whether one mean value is truly higher or lower than the other, I guess I want to know if they are only significantly different. Should I use two tail for this? If they are not significantly different, I can judge based on the mean values of test alone. And if they are significantly different, then I will need to do other type of analysis. Also, when I get my P-value for two tail, should I compare it to 0.025 or 0.05 if my confidence level is 0.05?

Thank you so much again.

July 31, 2020 at 11:19 pm

For your first, if you absolutely know that the mean must be lower the later the material is tested, that it cannot be higher, that would be a situation where you can use a one-tailed test. However, if that’s not a certainty, you’re just guessing, use a two-tail test. If you’re measuring different items at the different times, use the independent 2-sample t-test. However, if you’re measuring the same items at two time points, use the paired t-test. If it’s appropriate, using the paired t-test will give you more statistical power because it accounts for the variability between items. For more information, see my post about when it’s ok to use a one-tailed test .

For the mix of materials, use a two-tailed test because the effect truly can go either direction.

Always compare the p-value to your full significance level regardless of whether it’s a one or two-tailed test. Don’t divide the significance level in half.

' src=

June 17, 2020 at 2:56 pm

Is it possible that we reach to opposite conclusions if we use a critical value method and p value method Secondly if we perform one tail test and use p vale method to conclude our Ho, then do we need to convert sig value of 2 tail into sig value of one tail. That can be done just by dividing it with 2

June 18, 2020 at 5:17 pm

The p-value method and critical value method will always agree as long as you’re not changing anything about how the methodology.

If you’re using statistical software, you don’t need to make any adjustments. The software will do that for you.

However, if you calculating it by hand, you’ll need to take your significance level and then look in the table for your test statistic for a one-tailed test. For example, you’ll want to look up 5% for a one-tailed test rather than a two-tailed test. That’s not as simple as dividing by two. In this article, I show examples of one-tailed and two-tailed tests for the same degrees of freedom. The t critical value for the two-tailed test is +/- 2.086 while for the one-sided test it is 1.725. It is true that probability associated with those critical values doubles for the one-tailed test (2.5% -> 5%), but the critical value itself is not half (2.086 -> 1.725). Study the first several graphs in this article to see why that is true.

For the p-value, you can take a two-tailed p-value and divide by 2 to determine the one-sided p-value. However, if you’re using statistical software, it does that for you.

' src=

June 11, 2020 at 3:46 pm

Hello Jim, if you have the time I’d be grateful if you could shed some clarity on this scenario:

“A researcher believes that aromatherapy can relieve stress but wants to determine whether it can also enhance focus. To test this, the researcher selected a random sample of students to take an exam in which the average score in the general population is 77. Prior to the exam, these students studied individually in a small library room where a lavender scent was present. If students in this group scored significantly above the average score in general population [is this one-tailed or two-tailed hypothesis?], then this was taken as evidence that the lavender scent enhanced focus.”

Thank you for your time if you do decide to respond.

June 11, 2020 at 4:00 pm

It’s unclear from the information provided whether the researchers used a one-tailed or two-tailed test. It could be either. A two-tailed test can detect effects in both directions, so it could definitely detect an average group score above the population score. However, you could also detect that effect using a one-tailed test if it was set up correctly. So, there’s not enough information in what you provided to know for sure. It could be either.

However, that’s irrelevant to answering the question. The tricky part, as I see it, is that you’re not entirely sure about why the scores are higher. Are they higher because the lavender scent increased concentration or are they higher because the subjects have lower stress from the lavender? Or, maybe it’s not even related to the scent but some other characteristic of the room or testing conditions in which they took the test. You just know the scores are higher but not necessarily why they’re higher.

I’d say that, no, it’s not necessarily evidence that the lavender scent enhanced focus. There are competing explanations for why the scores are higher. Also, it would be best do this as an experiment with a control and treatment group where subjects are randomly assigned to either group. That process helps establish causality rather than just correlation and helps rules out competing explanations for why the scores are higher.

By the way, I spend a lot of time on these issues in my Introduction to Statistics ebook .

' src=

June 9, 2020 at 1:47 pm

If a left tail test has an alpha value of 0.05 how will you find the value in the table

' src=

April 19, 2020 at 10:35 am

Hi Jim, My question is in regards to the results in the table in your example of the one-sample T (Two-Tailed) test. above. What about the P-value? The P-value listed is .018. I assuming that is compared to and alpha of 0.025, correct?

In regression analysis, when I get a test statistic for the predictive variable of -2.099 and a p-value of 0.039. Am I comparing the p-value to an alpha of 0.025 or 0.05? Now if I run a Bootstrap for coefficients analysis, the results say the sig (2-tail) is 0.098. What are the critical values and alpha in this case? I’m trying to reconcile what I am seeing in both tables.

Thanks for your help.

April 20, 2020 at 3:24 am

Hi Marvalisa,

For one-tailed tests, you don’t need to divide alpha in half. If you can tell your software to perform a one-tailed test, it’ll do all the calculations necessary so you don’t need to adjust anything. So, if you’re using an alpha of 0.05 for a one-tailed test and your p-value is 0.04, it is significant. The procedures adjust the p-values automatically and it all works out. So, whether you’re using a one-tailed or two-tailed test, you always compare the p-value to the alpha with no need to adjust anything. The procedure does that for you!

The exception would be if for some reason your software doesn’t allow you to specify that you want to use a one-tailed test instead of a two-tailed test. Then, you divide the p-value from a two-tailed test in half to get the p-value for a one tailed test. You’d still compare it to your original alpha.

For regression, the same thing applies. If you want to use a one-tailed test for a cofficient, just divide the p-value in half if you can’t tell the software that you want a one-tailed test. The default is two-tailed. If your software has the option for one-tailed tests for any procedure, including regression, it’ll adjust the p-value for you. So, in the normal course of things, you won’t need to adjust anything.

' src=

March 26, 2020 at 12:00 pm

Hey Jim, for a one-tailed hypothesis test with a .05 confidence level, should I use a 95% confidence interval or a 90% confidence interval? Thanks

March 26, 2020 at 5:05 pm

You should use a one-sided 95% confidence interval. One-sided CIs have either an upper OR lower bound but remains unbounded on the other side.

' src=

March 16, 2020 at 4:30 pm

This is not applicable to the subject but… When performing tests of equivalence, we look at the confidence interval of the difference between two groups, and we perform two one-sided t-tests for equivalence..

' src=

March 15, 2020 at 7:51 am

Thanks for this illustrative blogpost. I had a question on one of your points though.

By definition of H1 and H0, a two-sided alternate hypothesis is that there is a difference in means between the test and control. Not that anything is ‘better’ or ‘worse’.

Just because we observed a negative result in your example, does not mean we can conclude it’s necessarily worse, but instead just ‘different’.

Therefore while it enables us to spot the fact that there may be differences between test and control, we cannot make claims about directional effects. So I struggle to see why they actually need to be used instead of one-sided tests.

What’s your take on this?

March 16, 2020 at 3:02 am

Hi Dominic,

If you’ll notice, I carefully avoid stating better or worse because in a general sense you’re right. However, given the context of a specific experiment, you can conclude whether a negative value is better or worse. As always in statistics, you have to use your subject-area knowledge to help interpret the results. In some cases, a negative value is a bad result. In other cases, it’s not. Use your subject-area knowledge!

I’m not sure why you think that you can’t make claims about directional effects? Of course you can!

As for why you shouldn’t use one-tailed tests for most cases, read my post When Can I Use One-Tailed Tests . That should answer your questions.

' src=

May 10, 2019 at 12:36 pm

Your website is absolutely amazing Jim, you seem like the nicest guy for doing this and I like how there’s no ulterior motive, (I wasn’t automatically signed up for emails or anything when leaving this comment). I study economics and found econometrics really difficult at first, but your website explains it so clearly its been a big asset to my studies, keep up the good work!

May 10, 2019 at 2:12 pm

Thank you so much, Jack. Your kind words mean a lot!

' src=

April 26, 2019 at 5:05 am

Hy Jim I really need your help now pls

One-tailed and two- tailed hypothesis, is it the same or twice, half or unrelated pls

April 26, 2019 at 11:41 am

Hi Anthony,

I describe how the hypotheses are different in this post. You’ll find your answers.

' src=

February 8, 2019 at 8:00 am

Thank you for your blog Jim, I have a Statistics exam soon and your articles let me understand a lot!

February 8, 2019 at 10:52 am

You’re very welcome! I’m happy to hear that it’s been helpful. Best of luck on your exam!

' src=

January 12, 2019 at 7:06 am

Hi Jim, When you say target value is 5. Do you mean to say the population mean is 5 and we are trying to validate it with the help of sample mean 4.1 using Hypo tests ?.. If it is so.. How can we measure a population parameter as 5 when it is almost impossible o measure a population parameter. Please clarify

January 12, 2019 at 6:57 pm

When you set a target for a one-sample test, it’s based on a value that is important to you. It’s not a population parameter or anything like that. The example in this post uses a case where we need parts that are stronger on average than a value of 5. We derive the value of 5 by using our subject area knowledge about what is required for a situation. Given our product knowledge for the hypothetical example, we know it should be 5 or higher. So, we use that in the hypothesis test and determine whether the population mean is greater than that target value.

When you perform a one-sample test, a target value is optional. If you don’t supply a target value, you simply obtain a confidence interval for the range of values that the parameter is likely to fall within. But, sometimes there is meaningful number that you want to test for specifically.

I hope that clarifies the rational behind the target value!

' src=

November 15, 2018 at 8:08 am

I understand that in Psychology a one tailed hypothesis is preferred. Is that so

November 15, 2018 at 11:30 am

No, there’s no overall preference for one-tailed hypothesis tests in statistics. That would be a study-by-study decision based on the types of possible effects. For more information about this decision, read my post: When Can I Use One-Tailed Tests?

' src=

November 6, 2018 at 1:14 am

I’m grateful to you for the explanations on One tail and Two tail hypothesis test. This opens my knowledge horizon beyond what an average statistics textbook can offer. Please include more examples in future posts. Thanks

November 5, 2018 at 10:20 am

Thank you. I will search it as well.

Stan Alekman

November 4, 2018 at 8:48 pm

Jim, what is the difference between the central and non-central t-distributions w/respect to hypothesis testing?

November 5, 2018 at 10:12 am

Hi Stan, this is something I will need to look into. I know central t-distribution is the common Student t-distribution, but I don’t have experience using non-central t-distributions. There might well be a blog post in that–after I learn more!

' src=

November 4, 2018 at 7:42 pm

this is awesome.

Comments and Questions Cancel reply

non directional two tailed hypothesis

  • The Open University
  • Accessibility hub
  • Guest user / Sign out
  • Study with The Open University

My OpenLearn Profile

Personalise your OpenLearn profile, save your favourite content and get recognition for your learning

About this free course

Become an ou student, download this course, share this free course.

Data analysis: hypothesis testing

Start this free course now. Just create an account and sign in. Enrol and complete the course for a free statement of participation or digital badge if available.

4.2 Two-tailed tests

Hypotheses that have an equal (=) or not equal (≠) supposition (sign) in the statement are called non-directional hypotheses . In non-directional hypotheses, the researcher is interested in whether there is a statistically significant difference or relationship between two or more variables, but does not have any specific expectation about which group or variable will be higher or lower. For example, a non-directional hypothesis might be: ‘There is a difference in the preference for brand X between male and female consumers.’ In this hypothesis, the researcher is interested in whether there is a statistically significant difference in the preference for brand X between male and female consumers, but does not have a specific prediction about which gender will have a higher preference. The researcher may conduct a survey or experiment to collect data on the brand preference of male and female consumers and then use statistical analysis to determine whether there is a significant difference between the two groups.

Non-directional hypotheses are also known as two-tailed hypotheses. The term ‘two-tailed’ comes from the fact that the statistical test used to evaluate the hypothesis is based on the assumption that the difference or relationship could occur in either direction, resulting in two ‘tails’ in the probability distribution. Using the coffee foam example (from Activity 1), you have the following set of hypotheses:

H 0 : µ = 1cm foam

H a : µ ≠ 1cm foam

In this case, the researcher can reject the null hypothesis for the mean value that is either ‘much higher’ or ‘much lower’ than 1 cm foam. This is called a two-tailed test because the rejection region includes outcomes from both the upper and lower tails of the sample distribution when determining a decision rule. To give an illustration, if you set alpha level (α) equal to 0.05, that would give you a 95% confidence level. Then, you would reject the null hypothesis for obtained values of z < 1.96 and z > 1.96 (you will look at how to calculate z-scores later in the course).

This can be plotted on a graph as shown in Figure 7.

A two-tailed test shown in a symmetrical graph reminiscent of a bell

A symmetrical graph reminiscent of a bell. The x-axis is labelled ‘z-score’ and the y-axis is labelled ‘probability density’. The x-axis increases in increments of 1 from -2 to 2.

The top of the bell-shaped curve is labelled ‘Foam height = 1cm’. The graph circles the rejection regions of the null hypothesis on both sides of the bell curve. Within these circles are two areas shaded orange: beneath the curve from -2 downwards which is labelled z < -1.96 and α = 0.025; and beneath the curve from 2 upwards which is labelled z > 1.96 and α = 0.025.

In a two-tailed hypothesis test, the null hypothesis assumes that there is no significant difference or relationship between the two groups or variables, and the alternative hypothesis suggests that there is a significant difference or relationship, but does not specify the direction of the difference or relationship.

When performing a two-tailed test, you need to determine the level of significance, which is denoted by alpha (α). The value of alpha, in this case, is 0.05. To perform a two-tailed test at a significance level of 0.05, you need to divide alpha by 2, giving a significance level of 0.025 for each distribution tail (0.05/2 = 0.025). This is done because the two-tailed test is looking for significance in either tail of the distribution. If the calculated test statistic falls in the rejection region of either tail of the distribution, then the null hypothesis is rejected and the alternative hypothesis is accepted. In this case, the researcher can conclude that there is a significant difference or relationship between the two groups or variables.

Assuming that the population follows a normal distribution, the tail located below the critical value of z = –1.96 (in a later section, you will discuss how this value was determined) and the tail above the critical value of z = +1.96 each represent a proportion of 0.025. These tails are referred to as the lower and upper tails, respectively, and they correspond to the extreme values of the distribution that are far from the central part of the bell curve. These critical values are used in a two-tailed hypothesis test to determine whether to reject or fail to reject the null hypothesis. The null hypothesis represents the default assumption that there is no significant difference between the observed data and what would be expected under a specific condition.

If the calculated test statistic falls within the critical values, then the null hypothesis cannot be rejected at the 0.05 level of significance. However, if the calculated test statistic falls outside the critical values (orange-coloured areas in Figure 7), then the null hypothesis can be rejected in favour of the alternative hypothesis, suggesting that there is evidence of a significant difference between the observed data and what would be expected under the specified condition.

Previous

Research Hypothesis In Psychology: Types, & Examples

Saul McLeod, PhD

Editor-in-Chief for Simply Psychology

BSc (Hons) Psychology, MRes, PhD, University of Manchester

Saul McLeod, PhD., is a qualified psychology teacher with over 18 years of experience in further and higher education. He has been published in peer-reviewed journals, including the Journal of Clinical Psychology.

Learn about our Editorial Process

Olivia Guy-Evans, MSc

Associate Editor for Simply Psychology

BSc (Hons) Psychology, MSc Psychology of Education

Olivia Guy-Evans is a writer and associate editor for Simply Psychology. She has previously worked in healthcare and educational sectors.

On This Page:

A research hypothesis, in its plural form “hypotheses,” is a specific, testable prediction about the anticipated results of a study, established at its outset. It is a key component of the scientific method .

Hypotheses connect theory to data and guide the research process towards expanding scientific understanding

Some key points about hypotheses:

  • A hypothesis expresses an expected pattern or relationship. It connects the variables under investigation.
  • It is stated in clear, precise terms before any data collection or analysis occurs. This makes the hypothesis testable.
  • A hypothesis must be falsifiable. It should be possible, even if unlikely in practice, to collect data that disconfirms rather than supports the hypothesis.
  • Hypotheses guide research. Scientists design studies to explicitly evaluate hypotheses about how nature works.
  • For a hypothesis to be valid, it must be testable against empirical evidence. The evidence can then confirm or disprove the testable predictions.
  • Hypotheses are informed by background knowledge and observation, but go beyond what is already known to propose an explanation of how or why something occurs.
Predictions typically arise from a thorough knowledge of the research literature, curiosity about real-world problems or implications, and integrating this to advance theory. They build on existing literature while providing new insight.

Types of Research Hypotheses

Alternative hypothesis.

The research hypothesis is often called the alternative or experimental hypothesis in experimental research.

It typically suggests a potential relationship between two key variables: the independent variable, which the researcher manipulates, and the dependent variable, which is measured based on those changes.

The alternative hypothesis states a relationship exists between the two variables being studied (one variable affects the other).

A hypothesis is a testable statement or prediction about the relationship between two or more variables. It is a key component of the scientific method. Some key points about hypotheses:

  • Important hypotheses lead to predictions that can be tested empirically. The evidence can then confirm or disprove the testable predictions.

In summary, a hypothesis is a precise, testable statement of what researchers expect to happen in a study and why. Hypotheses connect theory to data and guide the research process towards expanding scientific understanding.

An experimental hypothesis predicts what change(s) will occur in the dependent variable when the independent variable is manipulated.

It states that the results are not due to chance and are significant in supporting the theory being investigated.

The alternative hypothesis can be directional, indicating a specific direction of the effect, or non-directional, suggesting a difference without specifying its nature. It’s what researchers aim to support or demonstrate through their study.

Null Hypothesis

The null hypothesis states no relationship exists between the two variables being studied (one variable does not affect the other). There will be no changes in the dependent variable due to manipulating the independent variable.

It states results are due to chance and are not significant in supporting the idea being investigated.

The null hypothesis, positing no effect or relationship, is a foundational contrast to the research hypothesis in scientific inquiry. It establishes a baseline for statistical testing, promoting objectivity by initiating research from a neutral stance.

Many statistical methods are tailored to test the null hypothesis, determining the likelihood of observed results if no true effect exists.

This dual-hypothesis approach provides clarity, ensuring that research intentions are explicit, and fosters consistency across scientific studies, enhancing the standardization and interpretability of research outcomes.

Nondirectional Hypothesis

A non-directional hypothesis, also known as a two-tailed hypothesis, predicts that there is a difference or relationship between two variables but does not specify the direction of this relationship.

It merely indicates that a change or effect will occur without predicting which group will have higher or lower values.

For example, “There is a difference in performance between Group A and Group B” is a non-directional hypothesis.

Directional Hypothesis

A directional (one-tailed) hypothesis predicts the nature of the effect of the independent variable on the dependent variable. It predicts in which direction the change will take place. (i.e., greater, smaller, less, more)

It specifies whether one variable is greater, lesser, or different from another, rather than just indicating that there’s a difference without specifying its nature.

For example, “Exercise increases weight loss” is a directional hypothesis.

hypothesis

Falsifiability

The Falsification Principle, proposed by Karl Popper , is a way of demarcating science from non-science. It suggests that for a theory or hypothesis to be considered scientific, it must be testable and irrefutable.

Falsifiability emphasizes that scientific claims shouldn’t just be confirmable but should also have the potential to be proven wrong.

It means that there should exist some potential evidence or experiment that could prove the proposition false.

However many confirming instances exist for a theory, it only takes one counter observation to falsify it. For example, the hypothesis that “all swans are white,” can be falsified by observing a black swan.

For Popper, science should attempt to disprove a theory rather than attempt to continually provide evidence to support a research hypothesis.

Can a Hypothesis be Proven?

Hypotheses make probabilistic predictions. They state the expected outcome if a particular relationship exists. However, a study result supporting a hypothesis does not definitively prove it is true.

All studies have limitations. There may be unknown confounding factors or issues that limit the certainty of conclusions. Additional studies may yield different results.

In science, hypotheses can realistically only be supported with some degree of confidence, not proven. The process of science is to incrementally accumulate evidence for and against hypothesized relationships in an ongoing pursuit of better models and explanations that best fit the empirical data. But hypotheses remain open to revision and rejection if that is where the evidence leads.
  • Disproving a hypothesis is definitive. Solid disconfirmatory evidence will falsify a hypothesis and require altering or discarding it based on the evidence.
  • However, confirming evidence is always open to revision. Other explanations may account for the same results, and additional or contradictory evidence may emerge over time.

We can never 100% prove the alternative hypothesis. Instead, we see if we can disprove, or reject the null hypothesis.

If we reject the null hypothesis, this doesn’t mean that our alternative hypothesis is correct but does support the alternative/experimental hypothesis.

Upon analysis of the results, an alternative hypothesis can be rejected or supported, but it can never be proven to be correct. We must avoid any reference to results proving a theory as this implies 100% certainty, and there is always a chance that evidence may exist which could refute a theory.

How to Write a Hypothesis

  • Identify variables . The researcher manipulates the independent variable and the dependent variable is the measured outcome.
  • Operationalized the variables being investigated . Operationalization of a hypothesis refers to the process of making the variables physically measurable or testable, e.g. if you are about to study aggression, you might count the number of punches given by participants.
  • Decide on a direction for your prediction . If there is evidence in the literature to support a specific effect of the independent variable on the dependent variable, write a directional (one-tailed) hypothesis. If there are limited or ambiguous findings in the literature regarding the effect of the independent variable on the dependent variable, write a non-directional (two-tailed) hypothesis.
  • Make it Testable : Ensure your hypothesis can be tested through experimentation or observation. It should be possible to prove it false (principle of falsifiability).
  • Clear & concise language . A strong hypothesis is concise (typically one to two sentences long), and formulated using clear and straightforward language, ensuring it’s easily understood and testable.

Consider a hypothesis many teachers might subscribe to: students work better on Monday morning than on Friday afternoon (IV=Day, DV= Standard of work).

Now, if we decide to study this by giving the same group of students a lesson on a Monday morning and a Friday afternoon and then measuring their immediate recall of the material covered in each session, we would end up with the following:

  • The alternative hypothesis states that students will recall significantly more information on a Monday morning than on a Friday afternoon.
  • The null hypothesis states that there will be no significant difference in the amount recalled on a Monday morning compared to a Friday afternoon. Any difference will be due to chance or confounding factors.

More Examples

  • Memory : Participants exposed to classical music during study sessions will recall more items from a list than those who studied in silence.
  • Social Psychology : Individuals who frequently engage in social media use will report higher levels of perceived social isolation compared to those who use it infrequently.
  • Developmental Psychology : Children who engage in regular imaginative play have better problem-solving skills than those who don’t.
  • Clinical Psychology : Cognitive-behavioral therapy will be more effective in reducing symptoms of anxiety over a 6-month period compared to traditional talk therapy.
  • Cognitive Psychology : Individuals who multitask between various electronic devices will have shorter attention spans on focused tasks than those who single-task.
  • Health Psychology : Patients who practice mindfulness meditation will experience lower levels of chronic pain compared to those who don’t meditate.
  • Organizational Psychology : Employees in open-plan offices will report higher levels of stress than those in private offices.
  • Behavioral Psychology : Rats rewarded with food after pressing a lever will press it more frequently than rats who receive no reward.

Print Friendly, PDF & Email

non directional two tailed hypothesis

Hypothesis Testing for Means & Proportions

  •   1  
  • |   2  
  • |   3  
  • |   4  
  • |   5  
  • |   6  
  • |   7  
  • |   8  
  • |   9  
  • |   10  

On This Page sidebar

Hypothesis Testing: Upper-, Lower, and Two Tailed Tests

Type i and type ii errors.

Learn More sidebar

All Modules

More Resources sidebar

Z score Table

t score Table

The procedure for hypothesis testing is based on the ideas described above. Specifically, we set up competing hypotheses, select a random sample from the population of interest and compute summary statistics. We then determine whether the sample data supports the null or alternative hypotheses. The procedure can be broken down into the following five steps.  

  • Step 1. Set up hypotheses and select the level of significance α.

H 0 : Null hypothesis (no change, no difference);  

H 1 : Research hypothesis (investigator's belief); α =0.05

 

Upper-tailed, Lower-tailed, Two-tailed Tests

The research or alternative hypothesis can take one of three forms. An investigator might believe that the parameter has increased, decreased or changed. For example, an investigator might hypothesize:  

: μ > μ , where μ is the comparator or null value (e.g., μ =191 in our example about weight in men in 2006) and an increase is hypothesized - this type of test is called an ; : μ < μ , where a decrease is hypothesized and this is called a ; or : μ ≠ μ where a difference is hypothesized and this is called a .  

The exact form of the research hypothesis depends on the investigator's belief about the parameter of interest and whether it has possibly increased, decreased or is different from the null value. The research hypothesis is set up by the investigator before any data are collected.

 

  • Step 2. Select the appropriate test statistic.  

The test statistic is a single number that summarizes the sample information.   An example of a test statistic is the Z statistic computed as follows:

When the sample size is small, we will use t statistics (just as we did when constructing confidence intervals for small samples). As we present each scenario, alternative test statistics are provided along with conditions for their appropriate use.

  • Step 3.  Set up decision rule.  

The decision rule is a statement that tells under what circumstances to reject the null hypothesis. The decision rule is based on specific values of the test statistic (e.g., reject H 0 if Z > 1.645). The decision rule for a specific test depends on 3 factors: the research or alternative hypothesis, the test statistic and the level of significance. Each is discussed below.

  • The decision rule depends on whether an upper-tailed, lower-tailed, or two-tailed test is proposed. In an upper-tailed test the decision rule has investigators reject H 0 if the test statistic is larger than the critical value. In a lower-tailed test the decision rule has investigators reject H 0 if the test statistic is smaller than the critical value.  In a two-tailed test the decision rule has investigators reject H 0 if the test statistic is extreme, either larger than an upper critical value or smaller than a lower critical value.
  • The exact form of the test statistic is also important in determining the decision rule. If the test statistic follows the standard normal distribution (Z), then the decision rule will be based on the standard normal distribution. If the test statistic follows the t distribution, then the decision rule will be based on the t distribution. The appropriate critical value will be selected from the t distribution again depending on the specific alternative hypothesis and the level of significance.  
  • The third factor is the level of significance. The level of significance which is selected in Step 1 (e.g., α =0.05) dictates the critical value.   For example, in an upper tailed Z test, if α =0.05 then the critical value is Z=1.645.  

The following figures illustrate the rejection regions defined by the decision rule for upper-, lower- and two-tailed Z tests with α=0.05. Notice that the rejection regions are in the upper, lower and both tails of the curves, respectively. The decision rules are written below each figure.

Rejection Region for Upper-Tailed Z Test (H : μ > μ ) with α=0.05

The decision rule is: Reject H if Z 1.645.

 

 

α

Z

0.10

1.282

0.05

1.645

0.025

1.960

0.010

2.326

0.005

2.576

0.001

3.090

0.0001

3.719

Standard normal distribution with lower tail at -1.645 and alpha=0.05

Rejection Region for Lower-Tailed Z Test (H 1 : μ < μ 0 ) with α =0.05

The decision rule is: Reject H 0 if Z < 1.645.

a

Z

0.10

-1.282

0.05

-1.645

0.025

-1.960

0.010

-2.326

0.005

-2.576

0.001

-3.090

0.0001

-3.719

Standard normal distribution with two tails

Rejection Region for Two-Tailed Z Test (H 1 : μ ≠ μ 0 ) with α =0.05

The decision rule is: Reject H 0 if Z < -1.960 or if Z > 1.960.

0.20

1.282

0.10

1.645

0.05

1.960

0.010

2.576

0.001

3.291

0.0001

3.819

The complete table of critical values of Z for upper, lower and two-tailed tests can be found in the table of Z values to the right in "Other Resources."

Critical values of t for upper, lower and two-tailed tests can be found in the table of t values in "Other Resources."

  • Step 4. Compute the test statistic.  

Here we compute the test statistic by substituting the observed sample data into the test statistic identified in Step 2.

  • Step 5. Conclusion.  

The final conclusion is made by comparing the test statistic (which is a summary of the information observed in the sample) to the decision rule. The final conclusion will be either to reject the null hypothesis (because the sample data are very unlikely if the null hypothesis is true) or not to reject the null hypothesis (because the sample data are not very unlikely).  

If the null hypothesis is rejected, then an exact significance level is computed to describe the likelihood of observing the sample data assuming that the null hypothesis is true. The exact level of significance is called the p-value and it will be less than the chosen level of significance if we reject H 0 .

Statistical computing packages provide exact p-values as part of their standard output for hypothesis tests. In fact, when using a statistical computing package, the steps outlined about can be abbreviated. The hypotheses (step 1) should always be set up in advance of any analysis and the significance criterion should also be determined (e.g., α =0.05). Statistical computing packages will produce the test statistic (usually reporting the test statistic as t) and a p-value. The investigator can then determine statistical significance using the following: If p < α then reject H 0 .  

 

 

  • Step 1. Set up hypotheses and determine level of significance

H 0 : μ = 191 H 1 : μ > 191                 α =0.05

The research hypothesis is that weights have increased, and therefore an upper tailed test is used.

  • Step 2. Select the appropriate test statistic.

Because the sample size is large (n > 30) the appropriate test statistic is

  • Step 3. Set up decision rule.  

In this example, we are performing an upper tailed test (H 1 : μ> 191), with a Z test statistic and selected α =0.05.   Reject H 0 if Z > 1.645.

We now substitute the sample data into the formula for the test statistic identified in Step 2.  

We reject H 0 because 2.38 > 1.645. We have statistically significant evidence at a =0.05, to show that the mean weight in men in 2006 is more than 191 pounds. Because we rejected the null hypothesis, we now approximate the p-value which is the likelihood of observing the sample data if the null hypothesis is true. An alternative definition of the p-value is the smallest level of significance where we can still reject H 0 . In this example, we observed Z=2.38 and for α=0.05, the critical value was 1.645. Because 2.38 exceeded 1.645 we rejected H 0 . In our conclusion we reported a statistically significant increase in mean weight at a 5% level of significance. Using the table of critical values for upper tailed tests, we can approximate the p-value. If we select α=0.025, the critical value is 1.96, and we still reject H 0 because 2.38 > 1.960. If we select α=0.010 the critical value is 2.326, and we still reject H 0 because 2.38 > 2.326. However, if we select α=0.005, the critical value is 2.576, and we cannot reject H 0 because 2.38 < 2.576. Therefore, the smallest α where we still reject H 0 is 0.010. This is the p-value. A statistical computing package would produce a more precise p-value which would be in between 0.005 and 0.010. Here we are approximating the p-value and would report p < 0.010.                  

In all tests of hypothesis, there are two types of errors that can be committed. The first is called a Type I error and refers to the situation where we incorrectly reject H 0 when in fact it is true. This is also called a false positive result (as we incorrectly conclude that the research hypothesis is true when in fact it is not). When we run a test of hypothesis and decide to reject H 0 (e.g., because the test statistic exceeds the critical value in an upper tailed test) then either we make a correct decision because the research hypothesis is true or we commit a Type I error. The different conclusions are summarized in the table below. Note that we will never know whether the null hypothesis is really true or false (i.e., we will never know which row of the following table reflects reality).

Table - Conclusions in Test of Hypothesis

 

is True

Correct Decision

Type I Error

is False

Type II Error

Correct Decision

In the first step of the hypothesis test, we select a level of significance, α, and α= P(Type I error). Because we purposely select a small value for α, we control the probability of committing a Type I error. For example, if we select α=0.05, and our test tells us to reject H 0 , then there is a 5% probability that we commit a Type I error. Most investigators are very comfortable with this and are confident when rejecting H 0 that the research hypothesis is true (as it is the more likely scenario when we reject H 0 ).

When we run a test of hypothesis and decide not to reject H 0 (e.g., because the test statistic is below the critical value in an upper tailed test) then either we make a correct decision because the null hypothesis is true or we commit a Type II error. Beta (β) represents the probability of a Type II error and is defined as follows: β=P(Type II error) = P(Do not Reject H 0 | H 0 is false). Unfortunately, we cannot choose β to be small (e.g., 0.05) to control the probability of committing a Type II error because β depends on several factors including the sample size, α, and the research hypothesis. When we do not reject H 0 , it may be very likely that we are committing a Type II error (i.e., failing to reject H 0 when in fact it is false). Therefore, when tests are run and the null hypothesis is not rejected we often make a weak concluding statement allowing for the possibility that we might be committing a Type II error. If we do not reject H 0 , we conclude that we do not have significant evidence to show that H 1 is true. We do not conclude that H 0 is true.

Lightbulb icon signifying an important idea

 The most common reason for a Type II error is a small sample size.

return to top | previous page | next page

Content ©2017. All Rights Reserved. Date last modified: November 6, 2017. Wayne W. LaMorte, MD, PhD, MPH

Directional vs. Non-Directional Hypothesis in Research

In the world of research and statistical analysis, formulating hypotheses is a crucial step in the scientific process. Hypotheses guide researchers in making predictions and testing relationships between variables. When it comes to hypotheses, there are two main types: directional and non-directional.

Directional Hypothesis

A directional hypothesis, also known as a one-tailed hypothesis, is formulated with a specific predicted direction of the relationship between variables. It indicates an expectation of the relationship being either positive or negative.

Example of Directional Hypothesis

Advantages of directional hypothesis, non-directional hypothesis.

Non-directional hypotheses are often used when there is insufficient prior knowledge or theoretical basis to predict the direction of the relationship. It allows for a more exploratory approach, where the researcher is open to discovering the nature of the relationship through data analysis .

Example of Non-Directional Hypothesis

Read More: Population vs Sample | Examples

Advantages of Non-Directional Hypothesis:

Difference between directional and non-directional hypotheses.

Choosing Between Directional and Non-Directional Hypotheses: The choice between a directional and non-directional hypothesis depends on the research question, existing knowledge, and theoretical background. Here are a few considerations for selecting the appropriate type of hypothesis:

Directional vs. Non-Directional Hypothesis

AspectDirectional HypothesisNon-Directional Hypothesis
Specifies the expected direction of the effectDoes not specify the expected direction
One-tailed (focuses on one direction)Two-tailed (considers both positive and negative effects)
Often based on prior research or theoryMay lack prior knowledge or theoretical basis
Higher power to detect the specified directionPower is divided between both directions
Less flexible in exploring alternative outcomesMore flexible in considering different outcomes.
Higher confidence in the predicted directionEqual confidence in both positive and negative effects

Directional hypotheses offer specific predictions about the expected direction of the relationship, whereas non-directional hypotheses allow for more exploratory investigations without preconceived notions of the direction.

Remember, hypotheses serve as a roadmap for research, and regardless of their type, they play a crucial role in scientific inquiry and the pursuit of knowledge.

Other articles

Related posts, types of research questions, 8 types of validity in research | examples, operationalization of variables in research | examples | benefits, 10 types of variables in research: definitions and examples, what is a theoretical framework examples, types of reliability in research | example | ppt, independent vs. dependent variable, 4 research proposal examples | proposal sample, how to write a literature review in research: a step-by-step guide, how to write an abstract for a research paper, leave a reply cancel reply.

Hypothesis ( AQA A Level Psychology )

Revision note.

Claire Neeson

Psychology Content Creator

  • A hypothesis is a testable statement written as a prediction of what the researcher expects to find as a result of their experiment
  • A hypothesis should be no more than one sentence long
  • The hypothesis needs to include the independent variable (IV) and the dependent variable (DV)
  • For example - stating that you will measure ‘aggression’ is not enough ('aggression' has not been operationalised)
  • by exposing some children to an aggressive adult model whilst other children are not exposed to an aggressive adult model (operationalisation of the IV) 
  • number of imitative and non-imitative acts of aggression performed by the child (operationalisation of the DV)

The Experimental Hypothesis

  • Children who are exposed to an aggressive adult model will perform more acts of imitative and non-imitative aggression than children who have not been exposed to an aggressive adult model
  • The experimental hypothesis can be written as a  directional hypothesis or as a non-directional hypothesis

The Experimental Hypothesis: Directional 

  • A directional experimental hypothesis (also known as one-tailed)  predicts the direction of the change/difference (it anticipates more specifically what might happen)
  • A directional hypothesis is usually used when there is previous research which support a particular theory or outcome i.e. what a researcher might expect to happen
  • Participants who drink 200ml of an energy drink 5 minutes before running 100m will be faster (in seconds) than participants who drink 200ml of water 5 minutes before running 100m
  • Participants who learn a poem in a room in which loud music is playing will recall less of the poem's content than participants who learn the same poem in a silent room

 The Experimental Hypothesis: Non-Directional 

  • A non-directional experimental hypothesis (also known as two -tailed) does not predict the direction of the change/difference (it is an 'open goal' i.e. anything could happen)
  • A non-directional hypothesis is usually used when there is either no or little previous research which support a particular theory or outcome i.e. what the researcher cannot be confident as to what will happen
  • There will be a difference in time taken (in seconds) to run 100m depending on whether participants have drunk 200ml of an energy drink or 200ml of water 5 minutes before running 
  • There will be a difference in recall of a poem depending on whether participants learn the poem in a room in which loud music is playing or in a silent room

The Null Hypothesis

  • All published psychology research must include the null hypothesis
  • There will be no difference in children's acts of imitative and non-imitative aggression depending on whether they have observed an aggressive adult model or a non-aggressive adult model
  • The null hypothesis has to begin with the idea that the IV will have no effect on the DV  because until the experiment is run and the results are analysed it is impossible to state anything else! 
  • To put this in 'laymen's terms: if you bought a lottery ticket you could not predict that you are going to win the jackpot: you have to wait for the results to find out (spoiler alert: the chances of this happening are soooo low that you might as well save your cash!)
  • There will be no difference in time taken (in seconds) to run 100m depending on whether participants have drunk 200ml of an energy drink or 200ml of water 5 minutes before running 
  • There will be no difference in recall of a poem depending on whether participants learn the poem in a room in which loud music is playing or in a silent room
  • (NB this is not quite so slick and easy with a directional hypothesis as this sort of hypothesis will never begin with 'There will be a difference')
  • this is why the null hypothesis is so important - it tells the researcher whether or not their experiment has shown a difference in conditions (which is generally what they want to see, otherwise it's back to the drawing board...)

Worked example

Jim wants to test the theory that chocolate helps your ability to solve word-search puzzles

He believes that sugar helps memory as he has read some research on this in a text book

He puts up a poster in his sixth-form common room asking for people to take part after school one day and explains that they will be required to play two memory games, where eating chocolate will be involved

(a)  Should Jim use a directional hypothesis in this study? Explain your answer (2 marks)

(b)  Write a suitable hypothesis for this study. (4 marks)

a) Jim should use a directional hypothesis (1 mark)

    because previous research exists that states what might happen (2 nd mark)

b)  'Participants will remember more items from a shopping list in a memory game within the hour after eating 50g of chocolate, compared to when they have not consumed any chocolate'

  • 1 st mark for directional
  • 2 nd mark for IV- eating chocolate
  • 3 rd mark for DV- number of items remembered
  • 4 th mark for operationalising both IV & DV
  • If you write a non-directional or null hypothesis the mark is 0
  • If you do not get the direction correct the mark is zero
  • Remember to operationalise the IV & DV

You've read 0 of your 10 free revision notes

Unlock more, it's free, join the 100,000 + students that ❤️ save my exams.

the (exam) results speak for themselves:

Did this page help you?

  • Scientific Processes
  • Features of Science
  • Statistics: Inferential & Descriptive
  • Gender & Culture in Psychology
  • Free Will & Determinism
  • The Nature-Nurture Debate
  • Holism & Reductionism
  • Idiographic & Nomothetic Approaches
  • Ethical Implications of Research

Author: Claire Neeson

Claire has been teaching for 34 years, in the UK and overseas. She has taught GCSE, A-level and IB Psychology which has been a lot of fun and extremely exhausting! Claire is now a freelance Psychology teacher and content creator, producing textbooks, revision notes and (hopefully) exciting and interactive teaching materials for use in the classroom and for exam prep. Her passion (apart from Psychology of course) is roller skating and when she is not working (or watching 'Coronation Street') she can be found busting some impressive moves on her local roller rink.

non directional two tailed hypothesis

Reference Library

Collections

  • See what's new
  • All Resources
  • Student Resources
  • Assessment Resources
  • Teaching Resources
  • CPD Courses
  • Livestreams

Study notes, videos, interactive activities and more!

Psychology news, insights and enrichment

Currated collections of free resources

Browse resources by topic

  • All Psychology Resources

Resource Selections

Currated lists of resources

Non-Directional Hypothesis

A non-directional hypothesis is a two-tailed hypothesis that does not predict the direction of the difference or relationship (e.g. girls and boys are different in terms of helpfulness).

  • Share on Facebook
  • Share on Twitter
  • Share by Email

Research Methods: MCQ Revision Test 1 for AQA A Level Psychology

Topic Videos

Example Answers for Research Methods: A Level Psychology, Paper 2, June 2018 (AQA)

Exam Support

Example Answer for Question 14 Paper 2: AS Psychology, June 2017 (AQA)

Model answer for question 11 paper 2: as psychology, june 2016 (aqa), a level psychology topic quiz - research methods.

Quizzes & Activities

Our subjects

  • › Criminology
  • › Economics
  • › Geography
  • › Health & Social Care
  • › Psychology
  • › Sociology
  • › Teaching & learning resources
  • › Student revision workshops
  • › Online student courses
  • › CPD for teachers
  • › Livestreams
  • › Teaching jobs

Boston House, 214 High Street, Boston Spa, West Yorkshire, LS23 6AD Tel: 01937 848885

  • › Contact us
  • › Terms of use
  • › Privacy & cookies

© 2002-2024 Tutor2u Limited. Company Reg no: 04489574. VAT reg no 816865400.

Logo

One-tailed and Two-tailed Tests

by Karen Grace-Martin   30 Comments

I was recently asked about when to use one and two tailed tests.

The long answer is:  Use one tailed tests when you have a specific hypothesis about the direction of your relationship.  Some examples include you hypothesize that one group mean is larger than the other; you hypothesize that the correlation is positive; you hypothesize that the proportion is below .5.

The short answer is: Never use one tailed tests.

1. Only a few statistical tests even can have one tail: z tests and t tests.  So you’re severely limited.  F tests, Chi-square tests , etc. can’t accommodate one-tailed tests because their distributions are not symmetric.  Most statistical methods, such as regression and ANOVA , are based on these tests, so you will rarely have the chance to implement them.

2. Probably because they are rare, reviewers balk at one-tailed tests.  They tend to assume that you are trying to artificially boost the power of your test.  Theoretically, however, there is nothing wrong with them when the hypothesis and the statistical test are right for them.

Reader Interactions

' src=

June 10, 2022 at 5:48 am

Dear Karen,

I am running a regression analysis, I wonder if the significance in table output for ANOVA is inherently one-tailed or two-tailed? I have a directional hypothesis, so I am wondering if I still have to divide the significance value? Thank you in advance!

' src=

June 15, 2022 at 11:08 am

It’s two-tailed. And the F tests cannot be made directional.

' src=

October 9, 2019 at 7:46 pm

Say for example I am testing a hypothesis at .05 and it is a one tailed test. I expect the treatment to have a decrease and the cut off for significance is t = -1.67 so this is a left tail test. If t value after I run the test is +2.24 is this non significant since it is not in the hypothesized direction (negative)?

' src=

August 30, 2019 at 2:13 am

does a one-tailed test always require a one-sided (directional) hypothesis, and does a two-tailed test always require a two-sided (nondirectional) hypothesis?

Asked in the other direction: Does a one-sided (directional) hypothesis always require a one-sided test, and does a two-sided (nondirectional) hypothesis always require a two-sided test?

Thanks for clarifying Claudia

September 3, 2019 at 10:11 am

Hi Claudia,

Good question. So technically, when I am talking about one-tailed test, I do mean a test of a directional hypothesis. Likewise, I mean a non-directional hypothesis when I say two-tailed test.

In z-tests and t-tests, which are symmetric distributions, these terms are indeed interchangeable. But it definitely gets more complicated once you start talking about Chi-square tests and F-tests. Both of these have 0 at the left end and all high values of the statistic in one tail. So you can’t distinguish directions for these kinds of test statistics. So you can only test nondirectional hypotheses.

' src=

May 1, 2019 at 2:30 am

I have seen textbooks reporting confidence intervals of a standard deviation using low and high values of the chi-square statistic. Is it not based on a two-sided test?

' src=

August 28, 2018 at 7:14 am

There is no need to have a specific directional hypothesis (although you would usually have one), all that is needed to justify is to have a directional claim. Most claims in published scientific research and applied research are directional, since the moment you say the difference is positive or negative you have a directional claim. The only way to avoid it is to not mention the observed difference, or to state the difference as “plus or minus X” which would be ridiculous in most contexts.

Also, you should not mistake the tailed-ness of a statistical distribution with the tailed-ness of a hypothesis. While the Chi-Square or F-distribution might only have one tail, they can still be used for inference of one-sided and two-sided hypothesis alike. One can go as far back as Fisher and find examples of that (Statistical Methods for Research Workers).

If you are interested in more detailed arguments for the use of one-sided tests see the series of articles on The One-Sided Project website at https://www.onesided.org/ .

' src=

March 2, 2018 at 1:07 am

Hello, Good Morning to all, I am working on a project of Behavior Based Safety thesis work in which I have to assess data by SPSS software. Can any one could help me because i haven’t any idea about this software

May 17, 2018 at 10:06 am

Hi Muhammad, We have a number of free resources on SPSS, but if you’re in the middle of a project we have an on-demand tutorial that will get you not just started, but able to use it: http://theanalysisinstitute.com/introspss/

' src=

August 24, 2016 at 12:08 pm

Thank you for the discussions about the power for one-sided test. I agree that we should be careful when we decide to use a one-sided test. For Chi-square test when comparing two proportions, we can use two approaches: normal-theory method (the z-test) and contingency-table approach (the Chi-square test). For the normal-theory test, it requires a large sample size with n>5 or n*proportion >10. If your proposed study satisfied this requirement, we can use normal-theory method which is z-test to compute the power for the one-sided test. If you are interested in how to compute the one-sided test power, you can refer to the textbook by Marcello Pagano and Kimberlee Gauvreau’s 2nd edition titled “Principles of Biostatistics” section 14.5 on page 330 regarding sample size estimation for one-sided hypothesis test for proportions. Thanks.

' src=

September 1, 2015 at 2:30 am

Chi-squared tests are ALWAYS one-tailed… You only reject the null hypothesis when the test statistic falls into the right tail of the chi-square distribution.

What chi-squared tests are generally NOT is “directional”. They generally do not test whether the observed values are greater than or smaller than expected, only that they differ significantly.

' src=

January 26, 2016 at 10:57 pm

If “Chi-squared tests are ALWAYS one-tailed…”, why is Karen saying in her above statement that “F tests, Chi-square tests, etc. can’t accommodate one-tailed tests”?

Sorry, but I am bit confused. Can you help?

January 27, 2016 at 11:30 am

Hi Dippies,

Perhaps better wording is “F tests, Chi-square tests, etc. can’t accommodate directional tests.” Because there is only one tail for these distributions in which to find significance, it can’t distinguish between non-directional tests (eg, H1: mu1 – mu2 not equal to 0) and directional tests (eg, H1: mu1-mu2 greater than 0). In a t-test or z-test, we can either split alpha between two tails for a non-directional test or put alpha all into one tail for a directional test. We can then see whether we’re in the right tail based on the sign of the test statistic. F’s and Chi-sq values are squared. So they’re always positive. You get the same F value regardless of the direction of the means.

' src=

October 19, 2017 at 10:29 am

Thanks for the explanation… I am wondering: what happens if I have two methods, m1 and m2, and I want to show that m1 performs better (e.g. gives a higher value) than m2? Should I still use a two-tailed test? How can I show that m1 is, in fact, better (and not just different) than m2 (assuming the test proves significant)? Thanks for any comments.

' src=

July 22, 2015 at 6:14 pm

My hypothesis says there is a positive relation between religiosity and altruistic behavior. Would a two-tailed approach using Pearson Correlation do the trick?

Some very helpful information available here! Thank you.

' src=

July 1, 2015 at 2:50 am

Hi Karen, I run paired sample t-test and it just has P-two tailed. So how to convert P-two tailed to t critical one tailed. I have t critical two tailed and df already. Thanks so much

' src=

March 20, 2015 at 2:41 pm

I have used a one-tailed test but the effect went into the opposite direction. How do I have to calculate my p-Value now. Thanks for your reply, Mirco

' src=

February 3, 2015 at 10:24 pm

F-tests are almost always one-tailed. You would convert a two-tailed test’s p-value into a one-tailed test’s p-value by *halving* the p-value, not multiplying by 2 (as recommended above).

February 6, 2015 at 5:05 pm

Thanks, Paul. Yes. I fixed the half.

While it’s true that F-tests are one-tailed, they’re not testing directional hypotheses, the way a one-tailed t or z test does.

' src=

October 29, 2014 at 11:15 pm

I trying to complete the results of my study and need to know how to convert my 1-tailed results to 2-tailed? The company that ran my stats used a 1-tailed instead of 2-tailed, which as I understand is what I should have used to show directionality. Nicole

November 3, 2014 at 4:41 pm

Hi Nicole, just double them.

' src=

September 26, 2014 at 8:13 am

Very interesting. I am reading a much celebrated book (The weakness of Civil Society in Post-Communist Europe, by Marc Howard, 2003) in polticial science at the moment containing regression analysis with one tailed coeficients. This, togheter with that they only have a N of 23 (with 5 independent variables), raise my eyebrows. The results are also quite controversial….

What would you off-hand say about that?

September 26, 2014 at 12:27 pm

Without knowing anything else, the one tailed tests of coefficients wouldn’t worry me too much except for the fact you said it’s controversial. Which means perhaps the opposite result is reasonable.

The N of 23 is more of a concern. That’s pretty small. I find results like this are not bad per se. It’s fine to consider as one possible piece of information–they are great for spurring more research. But you can’t make any conclusions based on them.

' src=

June 14, 2013 at 9:54 pm

Jst wanna thank you for your post ; it will save me on my exam tomorrow (y)

& in my course it is not always divided by 2 (p-value). It depends on the value of your t (,= 0) & if your H1:’value” is > or 0, Hasard ?

Even if i’m wrong, I want to thank you again !

' src=

January 2, 2013 at 8:16 am

how can we calculate p-value of one-tailed from two-tailed hypothesis in spss?

January 2, 2013 at 10:50 am

All you have to do is divide it by 2.

' src=

December 1, 2012 at 11:04 am

i found this issue more important please continue in such way. kind regards, nurilign.

' src=

August 27, 2010 at 12:59 pm

Thank you so much for your reply and offer to try a find a source regarding the limited utility of one-tailed tests when doing ANOVAs and regressions, as well as the advice for converting two-tailed tests to one-tailed tests.

The only problem I have is that the Pearson Correlation Coefficient output from my stats consultant does not contain the p value. In order to calculate the p value for a two-tailed test, I thought it might be possible to take the df (n-2 for two-tailed tests) and look up the significance level in the table of critical values of the correlation coefficient. Once I get those values, I would simply divide by 2 to get the one-tailed level of significance. Do you think that is a statistically sound procedure.

Thank you again for your assistance, Sue

August 25, 2010 at 11:29 am

I am currently working on my dissertation and one of my committee members suggested that I should have used a one-tailed test as I have a directional hypothesis, but I think that a two-tailed test is just as appropriate based on several of the reasons listed on the blog.

I was particularly intrigued by the statement that “F tests, chi-square tests, etc. can’t accommodate one-tailed tests because their distributions are not symmetric.” This would make a fine argument for not re-rerunning my data and was wondering if there is a reference or citation for that point. I have not been able to find that point in any of the stats texts that I own. Any help would be greatly appreciated!

August 26, 2010 at 7:23 pm

Hi Sue, Hmmm. I would think that texts that talk about the F-test would mention that it's not symmetric. You could certainly use any text that states that t-squared=F. But I'll see if I can find something that says it directly. But in any case, you don't have to rerun anything, even if you weren't using F tests. To get a one-sided p-value, just halve the two-sided p-value you have. Karen

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Privacy Overview

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • Front Psychol

Fisher, Neyman-Pearson or NHST? A tutorial for teaching data testing

Despite frequent calls for the overhaul of null hypothesis significance testing (NHST), this controversial procedure remains ubiquitous in behavioral, social and biomedical teaching and research. Little change seems possible once the procedure becomes well ingrained in the minds and current practice of researchers; thus, the optimal opportunity for such change is at the time the procedure is taught, be this at undergraduate or at postgraduate levels. This paper presents a tutorial for the teaching of data testing procedures, often referred to as hypothesis testing theories. The first procedure introduced is Fisher's approach to data testing—tests of significance; the second is Neyman-Pearson's approach—tests of acceptance; the final procedure is the incongruent combination of the previous two theories into the current approach—NSHT. For those researchers sticking with the latter, two compromise solutions on how to improve NHST conclude the tutorial.

Introduction

This paper introduces the classic approaches for testing research data: tests of significance, which Fisher helped develop and promote starting in 1925; tests of statistical hypotheses, developed by Neyman and Pearson ( 1928 ); and null hypothesis significance testing (NHST), first concocted by Lindquist ( 1940 ). This chronological arrangement is fortuitous insofar it introduces the simpler testing approach by Fisher first, then moves onto the more complex one by Neyman and Pearson, before tackling the incongruent hybrid approach represented by NHST (Gigerenzer, 2004 ; Hubbard, 2004 ). Other theories, such as Bayes's hypotheses testing (Lindley, 1965 ) and Wald's ( 1950 ) decision theory, are not object of this tutorial.

The main aim of the tutorial is to illustrate the bases of discord in the debate against NHST (Macdonald, 2002 ; Gigerenzer, 2004 ), which remains a problem not only yet unresolved but very much ubiquitous in current data testing (e.g., Franco et al., 2014 ) and teaching (e.g., Dancey and Reidy, 2014 ), especially in the biological sciences (Lovell, 2013 ; Ludbrook, 2013 ), social sciences (Frick, 1996 ), psychology (Nickerson, 2000 ; Gigerenzer, 2004 ) and education (Carver, 1978 , 1993 ).

This tutorial is appropriate for the teaching of data testing at undergraduate and postgraduate levels, and is best introduced when students are knowledgeable on important background information regarding research methods (such as random sampling) and inferential statistics (such as frequency distributions of means).

In order to improve understanding, statistical constructs that may bring about confusion between theories are labeled differently, attending to their function in preference to their historical use (Perezgonzalez, 2014 ). Descriptive notes (notes) and caution notes (caution) are provided to clarify matters whenever appropriate.

Fisher's approach to data testing

Ronald Aylmer Fisher was the main force behind tests of significance (Neyman, 1967 ) and can be considered the most influential figure in the current approach to testing research data (Hubbard, 2004 ). Although some steps in Fisher's approach may be worked out a priori (e.g., the setting of hypotheses and levels of significance), the approach is eminently inferential and all steps can be set up a posteriori, once the research data are ready to be analyzed (Fisher, 1955 ; Macdonald, 1997 ). Some of these steps can even be omitted in practice, as it is relatively easy for a reader to recreate them. Fisher's approach to data testing can be summarized in the five steps described below.

Step 1–Select an appropriate test . This step calls for selecting a test appropriate to, primarily, the research goal of interest (Fisher, 1932 ), although you may also need to consider other issues, such as the way your variables have been measured. For example, if your research goal is to assess differences in the number of people in two independent groups, you would choose a chi-square test (it requires variables measured at nominal levels); on the other hand, if your interest is to assess differences in the scores that the people in those two groups have reported on a questionnaire, you would choose a t -test (it requires variables measured at interval or ratio levels and a close-to-normal distribution of the groups' differences).

Step 2–Set up the null hypothesis (H 0 ) . The null hypothesis derives naturally from the test selected in the form of an exact statistical hypothesis (e.g., H 0 : M1–M2 = 0; Neyman and Pearson, 1933 ; Carver, 1978 ; Frick, 1996 ). Some parameters of this hypothesis, such as variance and degrees of freedom, are estimated from the sample, while other parameters, such as the distribution of frequencies under a particular distribution, are deduced theoretically. The statistical distribution so established thus represents the random variability that is theoretically expected for a statistical nil hypothesis (i.e., H 0 = 0) given a particular research sample (Fisher, 1954 , 1955 ; Bakan, 1966 ; Macdonald, 2002 ; Hubbard, 2004 ). It is called the null hypothesis because it stands to be nullified with research data (Gigerenzer, 2004 ).

Among things to consider when setting the null hypothesis is its directionality.

Directional and non-directional hypotheses . With some research projects, the direction of the results is expected (e.g., one group will perform better than the other). In these cases, a directional null hypothesis covering all remaining possible results can be set (e.g., H 0 : M1–M2 = 0). With other projects, however, the direction of the results is not predictable or of no research interest. In these cases, a non-directional hypothesis is most suitable (e.g., H 0 : M1–M2 = 0).

H 0 does not need to be a nil hypothesis, that is, one that always equals zero (Fisher, 1955 ; Gigerenzer, 2004 ). For example, H 0 could be that the group difference is not larger than certain value (Newman et al., 2001 ). More often than not, however, H 0 tends to be zero.

Setting up H 0 is one of the steps usually omitted if following the typical nil expectation (e.g., no correlation between variables, no differences in variance among groups, etc.). Even directional nil hypotheses are often omitted, instead specifying that one-tailed tests (see below) have been used in the analysis.

Step 3-Calculate the theoretical probability of the results under H 0 ( p ) . Once the corresponding theoretical distribution is established, the probability ( p -value) of any datum under the null hypothesis is also established, which is what statistics calculate (Fisher, 1955 , 1960 ; Bakan, 1966 ; Johnstone, 1987 ; Cortina and Dunlap, 1997 ; Hagen, 1997 ). Data closer to the mean of the distribution (Figure ​ (Figure1) 1 ) have a greater probability of occurrence under the null distribution; that is, they appear more frequently and show a larger p -value (e.g., p = 0.46, or 46 times in a 100 trials). On the other hand, data located further away from the mean have a lower probability of occurrence under the null distribution; that is, they appear less often and, thus, show a smaller p -value (e.g., p = 0.003). Of interest to us is the probability of our research results under such null distribution (e.g., the probability of the difference in means between two research groups).

An external file that holds a picture, illustration, etc.
Object name is fpsyg-06-00223-g0001.jpg

Location of a t -value and its corresponding p -value on a theoretical t distribution with 30 degrees of freedom . The actual p -value conveys stronger evidence against H 0 than sig ≈0.05 and can be considered highly significant.

The p -value comprises the probability of the observed results and also of any other more extreme results (e.g., the probability of the actual difference between groups and any other difference more extreme than that). Thus, the p -value is a cumulative probability rather than an exact point probability: It covers the probability area extending from the observed results toward the tail of the distribution (Fisher, 1960 ; Carver, 1978 ; Frick, 1996 ; Hubbard, 2004 ).

P-values provide information about the theoretical probability of the observed and more extreme results under a null hypothesis assumed to be true (Fisher, 1960 ; Bakan, 1966 ), or, said otherwise, the probability of the data given a true hypothesis—P(D|H); (Carver, 1978 ; Hubbard, 2004 ). As H 0 is always true (i.e., it shows the theoretical random distribution of frequencies under certain parameters), it cannot, at the same time, be false nor falsifiable a posteriori. Basically, if at any point you say that H 0 is false, then you are also invalidating the whole test and its results. Furthermore, because H 0 is always true, it cannot be proved, either.

Step 4–Assess the statistical significance of the results . Fisher proposed tests of significance as a tool for identifying research results of interest, defined as those with a low probability of occurring as mere random variation of a null hypothesis. A research result with a low p -value may, thus, be taken as evidence against the null (i.e., as evidence that it may not explain those results satisfactorily; Fisher, 1960 ; Bakan, 1966 ; Johnstone, 1987 ; Macdonald, 2002 ). How small a result ought to be in order to be considered statistically significant is largely dependent on the researcher in question, and may vary from research to research (Fisher, 1960 ; Gigerenzer, 2004 ). The decision can also be left to the reader, so reporting exact p -values is very informative (Fisher, 1973 ; Macdonald, 1997 ; Gigerenzer, 2004 ).

Overall, however, the assessment of research results is largely made bound to a given level of significance, by comparing whether the research p -value is smaller than such level of significance or not (Fisher, 1954 , 1960 ; Johnstone, 1987 ):

  • If the p -value is approximately equal to or smaller than the level of significance, the result is considered statistically significant.
  • If the p -value is larger than the level of significance, the result is considered statistically non-significant.

Among things to consider when assessing the statistical significance of research results are the level of significance, and how it is affected by the directionality of the test and other corrections.

Level of significance (sig) . The level of significance is a theoretical p -value used as a point of reference to help identify statistically significant results (Figure ​ (Figure1). 1 ). There is no need to set up a level of significance a priori nor for a particular level of significance to be used in all occasions, although levels of significance such as 5% (sig ≈0.05) or 1% (sig ≈0.01) may be used for convenience, especially with novel research projects (Fisher, 1960 ; Carver, 1978 ; Gigerenzer, 2004 ). This highlights an important property of Fisher's levels of significance: They do not need to be rigid (e.g., p -values such as 0.049 and 0.051 have about the same statistical significance around a convenient level of significance of 5%; Johnstone, 1987 ).

Another property of tests of significance is that the observed p -value is taken as evidence against the null hypothesis, so that the smaller the p -value the stronger the evidence it provides (Fisher, 1960 ; Spielman, 1978 ). This means that it is plausible to gradate the strength of such evidence with smaller levels of significance. For example, if using 5% (sig ≈0.05) as a convenient level for identifying results which are just significant, then 1% (sig ≈0.01) may be used as a convenient level for identifying highly significant results and 1‰ (sig ≈0.001) for identifying extremely significant results.

Setting up a level of significance is another step usually omitted. In such cases, you may assume the researcher is using conventional levels of significance.

If both H 0 and sig are made explicit, they could be joined in a single postulate, such as H 0 : M 1 –M 2 = 0, sig ≈0.05.

Notice that the p-value informs about the probability associated with a given test value (e.g., a t value). You could use this test value to decide about the significance of your results in a fashion similar to Neyman-Pearson's approach (see below). However, you get more information about the strength of the research evidence with p-values.

Although the p-value is the most informative statistic of a test of significance, in psychology (e.g., American Psychological Association, 2010 ) you also report the research value of the test—e.g., t (30) = 2.25, p = 0.016, 1-tailed. Albeit cumbersome and largely ignored by the reader, the research value of the test offers potentially useful information (e.g., about the valid sample size used with a test).

Caution: Be careful not to interpret Fisher's p-values as Neyman-Pearson's Type I errors (α, see below). Probability values in single research projects are not the same than probability values in the long run (Johnstone, 1987 ), something illustrated by Berger ( 2003 )—who reported that p = 0.05 often corresponds to α = 0.5 (or anywhere between α = 0.22 and α > 0.5)—and Cumming ( 2014 )—who simulates the “dance” of p-values in the long run, commented further in Perezgonzalez ( 2015 ).

One-tailed and two-tailed tests . With some tests (e.g., F -tests) research data can only be tested against one side of the null distribution (one-tailed tests), while other tests (e.g., t -tests) can test research data against both sides of the null distribution at the same time. With one-tailed tests you set the level of significance on the appropriate tail of the distribution. With two-tailed tests you cover both eventualities by dividing the level of significance between both tails (Fisher, 1960 ; Macdonald, 1997 ), which is commonly done by halving the total level of significance in two equal areas (thus covering, for example, the 2.5% most extreme positive differences and the 2.5% most extreme negative differences).

The tail of a test depends on the test in question, not on whether the null hypothesis is directional or non-directional. However, you can use two-tailed tests as one-tailed ones when testing data against directional hypotheses.

Correction of the level of significance for multiple tests . As we introduced earlier, a p -value can be interpreted in terms of its expected frequency of occurrence under the specific null distribution for a particular test (e.g., p = 0.02 describes a result that is expected to appear 2 times out of 100 under H 0 ). The same goes for theoretical p -values used as levels of significance. Thus, if more than one test is performed, this has the consequence of also increasing the probability of finding statistical significant results which are due to mere chance variation. In order to keep such probability at acceptable levels overall, the level of significance may be corrected downwards (Hagen, 1997 ). A popular correction is Bonferroni's, which reduces the level of significance proportionally to the number of tests carried out. For example, if your selected level of significance is 5% (sig ≈0.05) and you carry out two tests, then such level of significance is maintained overall by correcting the level of significance for each test down to 2.5% (sig ≈0.05/2 tests ≈0.025, or 2.5% per test).

Bonferroni's correction is popular but controversial, mainly because it is too conservative, more so as the number of multiple tests increases. There are other methods for controlling the probability of false results when doing multiple comparisons, including familywise error rate methods (e.g., Holland and Copenhaver, 1987 ), false discovery rate methods (e.g., Benjamini and Hochberg, 1995 ), resampling methods (jackknifing, bootstrapping— e.g., Efron, 1981 ), and permutation tests (i.e., exact tests—e.g., Gill, 2007 ).

Step 5–Interpret the statistical significance of the results . A significant result is literally interpreted as a dual statement: Either a rare result that occurs only with probability p (or lower) just happened, or the null hypothesis does not explain the research results satisfactorily (Fisher, 1955 ; Carver, 1978 ; Johnstone, 1987 ; Macdonald, 1997 ). Such literal interpretation is rarely encountered, however, and most common interpretations are in the line of “The null hypothesis did not seem to explain the research results well, thus we inferred that other processes—which we believe to be our experimental manipulation—exist that account for the results,” or “The research results were statistically significant, thus we inferred that the treatment used accounted for such difference.”

Non-significant results may be ignored (Fisher, 1960 ; Nunnally, 1960 ), although they can still provide useful information, such as whether results were in the expected direction and about their magnitude (Fisher, 1955 ). In fact, although always denying that the null hypothesis could ever be supported or established, Fisher conceded that non-significant results might be used for confirming or strengthening it (Fisher, 1955 ; Johnstone, 1987 ).

Statistically speaking, Fisher's approach only ascertains the probability of the research data under a null hypothesis. Doubting or denying such hypothesis given a low p-value does not necessarily “support” or “prove” that the opposite is true (e.g., that there is a difference or a correlation in the population). More importantly, it does not “support” or “prove” that whatever else has been done in the research (e.g., the treatment used) explains the results, either (Macdonald, 1997 ). For Fisher, a good control of the research design (Fisher, 1955 ; Johnstone, 1987 ; Cortina and Dunlap, 1997 ), especially random allocation, is paramount to make sensible inferences based on the results of tests of significance (Fisher, 1954 ; Neyman, 1967 ). He was also adamant that, given a significant result, further research was needed to establish that there has indeed been an effect due to the treatment used (Fisher, 1954 ; Johnstone, 1987 ; Macdonald, 2002 ). Finally, he considered significant results as mere data points and encouraged the use of meta-analysis for progressing further, combining significant and non-significant results from related research projects (Fisher, 1960 ; Neyman, 1967 ).

Highlights of Fisher's approach

Flexibility . Because most of the work is done a posteriori, Fisher's approach is quite flexible, allowing for any number of tests to be carried out and, therefore, any number of null hypotheses to be tested (a correction of the level of significance may be appropriate, though—Macdonald, 1997 ).

Better suited for ad-hoc research projects . Given above flexibility, Fisher's approach is well suited for single, ad-hoc , research projects (Neyman, 1956 ; Johnstone, 1987 ), as well as for exploratory research (Frick, 1996 ; Macdonald, 1997 ; Gigerenzer, 2004 ).

Inferential . Fisher's procedure is largely inferential, from the sample to the population of reference, albeit of limited reach, mainly restricted to populations that share parameters similar to those estimated from the sample (Fisher, 1954 , 1955 ; Macdonald, 2002 ; Hubbard, 2004 ).

No power analysis . Neyman ( 1967 ) and Kruskal and Savage (Kruskal, 1980 ) were surprised that Fisher did not explicitly attend to the power of a test. Fisher talked about sensitiveness, a similar concept, and how it could be increased by increasing sample size (Fisher, 1960 ). However, he never created a mathematical procedure for controlling sensitiveness in a predictable manner (Macdonald, 1997 ; Hubbard, 2004 ).

No alternative hypothesis . One of the main critiques to Fisher's approach is the lack of an explicit alternative hypothesis (Macdonald, 2002 ; Gigerenzer, 2004 ; Hubbard, 2004 ), because there is no point in rejecting a null hypothesis without an alternative explanation being available (Pearson, 1990 ). However, Fisher considered alternative hypotheses implicitly—these being the negation of the null hypotheses—so much so that for him the main task of the researcher—and a definition of a research project well done—was to systematically reject with enough evidence the corresponding null hypothesis (Fisher, 1960 ).

Neyman-Pearson's approach to data testing

Jerzy Neyman and Egon Sharpe Pearson tried to improve Fisher's procedure (Fisher, 1955 ; Pearson, 1955 ; Jones and Tukey, 2000 ; Macdonald, 2002 ) and ended up developing an alternative approach to data testing. Neyman-Pearson's approach is more mathematical than Fisher's and does much of its work a priori, at the planning stage of the research project (Fisher, 1955 ; Macdonald, 1997 ; Gigerenzer, 2004 ; Hubbard, 2004 ). It also introduces a number of constructs, some of which are similar to those of Fisher. Overall, Neyman-Pearson's approach to data testing can be considered tests of acceptance (Fisher, 1955 ; Pearson, 1955 ; Spielman, 1978 ; Perezgonzalez, 2014 ), summarized in the following eight main steps.

A priori steps

Step 1–Set up the expected effect size in the population . The main conceptual innovation of Neyman-Pearson's approach was the consideration of explicit alternative hypotheses when testing research data (Neyman and Pearson, 1928 , 1933 ; Neyman, 1956 ; Macdonald, 2002 ; Gigerenzer, 2004 ; Hubbard, 2004 ). In their simplest postulate, the alternative hypothesis represents a second population that sits alongside the population of the main hypothesis on the same continuum of values. These two groups differ by some degree: the effect size (Cohen, 1988 ; Macdonald, 1997 ).

Although the effect size was a new concept introduced by Neyman and Pearson, in psychology it was popularized by Cohen ( 1988 ). For example, Cohen's conventions for capturing differences between groups—d (Figure ​ (Figure2)—were 2 )—were based on the degree of visibility of such differences in the population: the smaller the effect size, the more difficult to appreciate such differences; the larger the effect size, the easier to appreciate such differences. Thus, effect sizes also double as a measure of importance in the real world (Nunnally, 1960 ; Cohen, 1988 ; Frick, 1996 ).

An external file that holds a picture, illustration, etc.
Object name is fpsyg-06-00223-g0002.jpg

A conventional large difference—Cohen's d = 0.8 —between two normally distributed populations, as a fraction of one standard deviation .

When testing data about samples, however, statistics do not work with unknown population distributions but with distributions of samples, which have narrower standard errors. In these cases, the effect size can still be defined as above because the means of the populations remain unaffected, but the sampling distributions would appear separated rather than overlapping (Figure ​ (Figure3). 3 ). Because we rarely know the parameters of populations, it is their equivalent effect size measures in the context of sampling distributions which are of interest.

An external file that holds a picture, illustration, etc.
Object name is fpsyg-06-00223-g0003.jpg

Sampling distributions ( N = 50 each) of the populations in Figure ​ Figure2 2 . MES ( d = 0.8), assuming β = 0.20, is d = 0.32 (i.e., the expected difference in the population ranges between d = 0.32 and infinity).

As we shall see below, the alternative hypothesis is the one that provides information about the effect size to be expected. However, because this hypothesis is not tested, Neyman-Pearson's procedure largely ignores its distribution except for a small percentage of it, which is called “beta” (β; Gigerenzer, 2004 ). Therefore, it is easier to understand Neyman-Pearson's procedure if we peg the effect size to beta and call it the expected minimum effect size (MES; Figure ​ Figure3). 3 ). This helps us conceptualize better how Neyman-Pearson's procedure works (Schmidt, 1996 ): The minimum effect size effectively represents that part of the main hypothesis that is not going to be rejected by the test ( i.e., MES captures values of no research interest which you want to leave under H M ; Cortina and Dunlap, 1997 ; Hagen, 1997 ; Macdonald, 2002 ). (Worry not, as there is no need to perform any further calculations: The population effect size is the one to use, for example, for estimating research power.)

A particularity of Neyman-Pearson's approach is that the two hypotheses are assumed to represent defined populations, the research sample being an instance of either of them (i.e., they are populations of samples generated by repetition of a common random process—Neyman and Pearson, 1928 ; Pearson, 1955 ; Hagen, 1997 ; Hubbard, 2004 ). This is unlike Fisher's population, which can be considered more theoretical, generated ad-hoc so as for providing the appropriate random distribution for the research sample at hand (i.e., a population of samples similar to the research sample—Fisher, 1955 ; Johnstone, 1987 ).

Step 2–Select an optimal test . As we shall see below, another of Neyman-Pearson's contributions was the construct of the power of a test. A spin-off of this contribution is that it has been possible to establish which tests are most powerful (for example, parametric tests are more powerful than non-parametric tests, and one-tailed tests are more powerful than two-tailed tests), and under which conditions (for example, increasing sample size increases power). For Neyman and Pearson, thus, you are better off choosing the most powerful test for your research project (Neyman, 1942 , 1956 ).

Step 3–Set up the main hypothesis (H M ) . Neyman-Pearson's approach considers, at least, two competing hypotheses, although it only tests data under one of them. The hypothesis which is the most important for the research (i.e., the one you do not want to reject too often) is the one tested (Neyman and Pearson, 1928 ; Neyman, 1942 ; Spielman, 1973 ). This hypothesis is better off written so as for incorporating the minimum expected effect size within its postulate (e.g., H M : M 1 –M 2 = 0 ± MES), so that it is clear that values within such minimum threshold are considered reasonably probable under the main hypothesis, while values outside such minimum threshold are considered as more probable under the alternative hypothesis (Figure ​ (Figure4 4 ).

An external file that holds a picture, illustration, etc.
Object name is fpsyg-06-00223-g0004.jpg

Neyman-Pearson's approach tests data under H M using the rejection region delimited by α . H A contributes MES and β. Differences of research interest will be equal or larger than MES and will fall within this rejection region.

Neyman-Pearson's H M is very similar to Fisher's H 0 . Indeed, Neyman and Pearson also called it the null hypothesis and often postulated it in a similar manner (e.g., as H M : M 1 –M 2 = 0). However, this similarity is merely superficial on three accounts: H M needs to be considered at the design stage (H 0 is rarely made explicit); it is implicitly designed to incorporate any value below the MES—i.e., the a priori power analysis of a test aims to capture such minimum difference (effect sizes are not part of Fisher's approach); and it is but one of two competing explanations for the research results (H 0 is the only hypothesis, to be nullified with evidence).

The main aspect to consider when setting the main hypothesis is the Type I error you want to control for during the research.

Type I error . A Type I error (or error of the first class) is made every time the main hypothesis is wrongly rejected (thus, every time the alternative hypothesis is wrongly accepted). Because the hypothesis under test is your main hypothesis, this is an error that you want to minimize as much as possible in your lifetime research (Neyman and Pearson, 1928 , 1933 ; Neyman, 1942 ; Macdonald, 1997 ).

A Type I error is possible under Fisher's approach, as it is similar to the error made when rejecting H 0 (Carver, 1978 ). However, this similarity is merely superficial on two accounts: Neyman and Pearson considered it an error whose relevance only manifests itself in the long run because it is not possible to know whether such an error has been made in any particular trial (Fisher's approach is eminently ad-hoc, so the risk of a long-run Type I error is of little relevance); therefore, it is an error that needs to be considered and minimized at the design stage of the research project in order to ensure good power—you cannot minimize this error a posteriori (with Fisher's approach, the potential impact of errors on individual projects is better controlled by correcting the level of significance as appropriate, for example, with a Bonferroni correction).

Alpha (α) . Alpha is the probability of committing a Type I error in the long run (Gigerenzer, 2004 ). Neyman and Pearson often worked with convenient alpha levels such as 5% (α = 0.05) and 1% (α = 0.01), although different levels can also be set. The main hypothesis can, thus, be written so as for incorporating the alpha level in its postulate (e.g., H M : M 1 –M 2 = 0 ± MES, α = 0.05), to be read as the probability level at which the main hypothesis will be rejected in favor of the alternative hypothesis.

Neyman-Pearson's α looks very similar to Fisher's sig. Indeed, Neyman and Pearson also called it the significance level of the test and used the same conventional cut-off points (5, 1%). However, this similarity is merely superficial on three accounts: α needs to be set a priori (not necessarily so under Fisher's approach); Neyman-Pearson's approach is not a test of significance (they are not interested in the strength of the evidence against H M ) but a test of acceptance (deciding whether to accept H A instead of H M ); and α does not admit gradation—i.e., you may choose, for example, either α = 0.05 or α = 0.01, but not both, for the same test (while with Fisher's approach you can have different levels of more extreme significance).

The critical region (CR test ) and critical value (CV test , Test crit ) of a test . The alpha level helps draw a critical region, or rejection region (Figure ​ (Figure4), 4 ), on the probability distribution of the main hypothesis (Neyman and Pearson, 1928 ). Any research value that falls outside this critical region will be taken as reasonably probable under the main hypothesis, and any research result that falls within the critical region will be taken as most probable under the alternative hypothesis. The alpha level, thus, also helps identify the location of the critical value of such test, the boundary for deciding between hypotheses. Thus, once the critical value is known—see below—, the main hypothesis can also be written so as for incorporating such critical value, if so desired (e.g., H M : M 1 –M 2 = 0 ± MES, α = 0.05, CV t = 2.38).

Neyman-Pearson's critical region is very similar to the equivalent critical region you would obtain by using Fisher's sig as a cut-off point on a null distribution. However, this similarity is rather unimportant on three accounts: it is based on a critical value which delimits the region to reject H M in favor of H A , irrespective of the actual observed value of the test (Fisher, on the contrary, is more interested in the actual p-value of the research result); it is fixed a priori and, thus, rigid and immobile (Fisher's level of significance can be flexible—Macdonald, 2002 ); and it is non-gradable (with Fisher's approach, you may delimit several more extreme critical regions as areas of stronger evidence).

Step 4–Set up the alternative hypothesis (H A ) . One of the main innovations of Neyman-Pearson's approach was the consideration of alternative hypotheses (Neyman and Pearson, 1928 , 1933 ; Neyman, 1956 ). Unfortunately, the alternative hypothesis is often postulated in an unspecified manner (e.g., as H A : M 1 –M 2 ≠ 0), even by Neyman and Pearson themselves (Macdonald, 1997 ; Jones and Tukey, 2000 ). In practice, a fully specified alternative hypothesis (e.g., its mean and variance) is not necessary because this hypothesis only provides partial information to the testing of the main hypothesis (a.k.a., the effect size and β). Therefore, the alternative hypothesis is better written so as for incorporating the minimum effect size within its postulate (e.g., H A : M 1 –M 2 ≠ 0 ± MES). This way it is clear that values beyond such minimum effect size are the ones considered of research importance.

Neyman-Pearson's H A is often postulated as the negation of a nil hypothesis (H A : M 1 –M 2 ≠ 0), which is coherent with a simple postulate of H M (H M : M 1 –M 2 = 0). These simplified postulates are not accurate and are easily confused with Fisher's approach to data testing—H M resembles Fisher's H 0 , and H A resembles a mere negation of H 0 . However, merely negating H 0 does not make its negation a valid alternative hypothesis—otherwise Fisher would have put forward such alternative hypothesis, something which he was vehemently against (Hubbard, 2004 ). As discussed earlier, Neyman-Pearson's approach introduces the construct of effect size into their testing approach; thus, incorporating such construct in the specification of both H M and H A makes them more accurate, and less confusing, than their simplified versions.

Among things to consider when setting the alternative hypothesis are the expected effect size in the population (see above) and the Type II error you are prepared to commit.

Type II error . A Type II error (or error of the second class) is made every time the main hypothesis is wrongly retained (thus, every time H A is wrongly rejected). Making a Type II error is less critical than making a Type I error, yet you still want to minimize the probability of making this error once you have decided which alpha level to use (Neyman and Pearson, 1933 ; Neyman, 1942 ; Macdonald, 2002 ).

Beta (β) . Beta is the probability of committing a Type II error in the long run and is, therefore, a parameter of the alternative hypothesis (Figure ​ (Figure4, 4 , Neyman, 1956 ). You want to make beta as small as possible, although not smaller than alpha (if β needed to be smaller than α, then H A should be your main hypothesis, instead!). Neyman and Pearson proposed 20% (β = 0.20) as an upper ceiling for beta, and the value of alpha (β = α) as its lower floor (Neyman, 1953 ). For symmetry with the main hypothesis, the alternative hypothesis can, thus, be written so as for incorporating the beta level in its postulate (e.g., H A : M 1 –M 2 ≠ 0 ± MES, β = 0.20).

Step 5–Calculate the sample size (N) required for good power (1–β) . Neyman-Pearson's approach is eminently a priori in order to ensure that the research to be done has good power (Neyman, 1942 , 1956 ; Pearson, 1955 ; Macdonald, 2002 ). Power is the probability of correctly rejecting the main hypothesis in favor of the alternative hypothesis (i.e., of correctly accepting H A ). It is the mathematical opposite of the Type II error (thus, 1–β; Macdonald, 1997 ; Hubbard, 2004 ). Power depends on the type of test selected (e.g., parametric tests and one-tailed tests increase power), as well as on the expected effect size (larger ES's increase power), alpha (larger α's increase power) and beta (smaller β 's increase power). A priori power is ensured by calculating the correct sample size given those parameters (Spielman, 1973 ). Because power is the opposite of beta, the lower floor for good power is, thus, 80% (1–β = 0.80), and its upper ceiling is 1–alpha (1–β = 1–α).

H A does not need to be tested under Neyman-Pearson's approach, only H M (Neyman and Pearson, 1928 , 1933 ; Neyman, 1942 ; Pearson, 1955 ; Spielman, 1973 ). Therefore, the procedure looks similar to Fisher's and, under similar circumstances (e.g., when using the same test and sample size), it will lead to the same results. The main difference between procedures is that Neyman-Pearson's H A provides explicit information to the test; that is, information about ES and β. If this information is not taken into account for designing a research project with adequate power, then, by default, you are carrying out a test under Fisher's approach.

Caution: For Neyman and Pearson, there is little justification in carrying out research projects with low power. When a research project has low power, Type II errors are too big, so it is less probable to reject H M in favor of H A , while, at the same time, it makes unreasonable to accept H M as the best explanation for the research results. If you face a research project with low a priori power, try the best compromise between its parameters (such as increasing α, relaxing β, settling for a larger ES, or using one-tailed tests; Neyman and Pearson, 1933 ). If all fails, consider Fisher's approach, instead.

Step 6–Calculate the critical value of the test (CV test , or Test crit ) . Some of above parameters (test, α and N) can be used for calculating the critical value of the test; that is, the value to be used as the cut-off point for deciding between hypotheses (Figure ​ (Figure5, 5 , Neyman and Pearson, 1933 ).

An external file that holds a picture, illustration, etc.
Object name is fpsyg-06-00223-g0005.jpg

Neyman-Pearson's test in action: CV test is the point for deciding between hypotheses; it coincides with the cut-off points underlying α, β, and MES .

A posteriori steps

Step 7–Calculate the test value for the research (RV test ) . In order to carry out the test, some unknown parameters of the populations are estimated from the sample (e.g., variance), while other parameters are deduced theoretically (e.g., the distribution of frequencies under a particular statistical distribution). The statistical distribution so established thus represents the random variability that is theoretically expected for a statistical main hypothesis given a particular research sample, and provides information about the values expected at different locations under such distribution.

By applying the corresponding formula, the research value of the test (RV test ) is obtained. This value is closer to zero the closer the research data is to the mean of the main hypothesis; it gets larger the further away the research data is from the mean of the main hypothesis.

P-values can also be used for testing data when using Neyman-Pearson's approach, as testing data under H M is similar to testing data under Fisher's H 0 (Fisher, 1955 ). It implies calculating the theoretical probability of the research data under the distribution of H M — P ( D | H M ). Just be mindful that p-values go in the opposite way than RVs, with larger p-values being closer to H M and smaller p-values being further away from it.

Caution: Because of above equivalence, you may use p-values instead of CV test with Neyman-Pearson's approach. However, p-values need to be considered mere proxies under this approach and, thus, have no evidential properties whatsoever (Frick, 1996 ; Gigerenzer, 2004 ). For example, if working with a priori α = 0.05, p = 0.01 would lead you to reject H M at α = 0.05; however, it would be incorrect to reject it at α = 0.01 (i.e., α cannot be adjusted a posteriori), and it would be incorrect to conclude that you reject H M strongly (i.e., α cannot be gradated). If confused, you are better off sticking to CV test , and using p-values only with Fisher's approach.

Step 8–Decide in favor of either the main or the alternative hypothesis . Neyman-Pearson's approach is rather mechanical once the a priori steps have been satisfied (Neyman and Pearson, 1933 ; Neyman, 1942 , 1956 ; Spielman, 1978 ; Macdonald, 2002 ). Thus, the analysis is carried out as per the optimal test selected and the interpretation of results is informed by the mathematics of the test, following on the a priori pattern set up for deciding between hypotheses:

  • If the observed result falls within the critical region, reject the main hypothesis and accept the alternative hypothesis.
  • If the observed result falls outside the critical region and the test has good power, accept the main hypothesis.
  • If the observed result falls outside the critical region and the test has low power, conclude nothing. (Ideally, you would not carry out research with low power—Neyman, 1955 ).

Neyman-Pearson's approach leads to a decision between hypotheses (Neyman and Pearson, 1933 ; Spielman, 1978 ). In principle, this decision should be between rejecting H M or retaining H M (assuming good power), as the test is carried out on H M only (Neyman, 1942 ). In practice, it does not really make much difference whether you accept H M or H A , as appropriate (Macdonald, 1997 ). In fact, accepting either H M or H A is beneficial as it prevents confusion with Fisher's approach, which can only reject H 0 (Perezgonzalez, 2014 ).

Reporting the observed research test value is relevant under Neyman-Pearson's approach, as it serves to compare the observed value against the a priori critical value—e.g., t (64) = 3.31, 1-tailed > CV t = 2.38, thus accept H A . When using a p-value as a proxy for CV test , simply strip any evidential value off p—e.g., t (64) = 3.31, p < α, 1-tailed.

Neyman-Pearson's hypotheses are also assumed to be true. H M represents the probability distribution of the data given a true hypothesis—P(D|H M ), while H A represents the distribution of the data under an alternative true hypothesis—P(D|H A ), even when it is never tested. This means that H M and H A cannot be, at the same time false, nor proved or falsified a posteriori. The only way forward is to act as if the conclusion reached by the test was true—subject to a probability α or β of making a Type I or Type II error, respectively (Neyman and Pearson, 1933 ; Cortina and Dunlap, 1997 ).

Highlights of Neyman-Pearson's approach

More powerful . Neyman-Pearson's approach is more powerful than Fisher's for testing data in the long run (Williams et al., 2006 ). However, repeated sampling is rare in research (Fisher, 1955 ).

Better suited for repeated sampling projects . Because of above, Neyman-Pearson's approach is well-suited for repeated sampling research using the same population and tests, such as industrial quality control or large scale diagnostic testing (Fisher, 1955 ; Spielman, 1973 ).

Deductive . The approach is deductive and rather mechanical once the a priori steps have been set up (Neyman and Pearson, 1933 ; Neyman, 1942 ; Fisher, 1955 ).

Less flexible than Fisher's approach . Because most of the work is done a priori, this approach is less flexible for accommodating tests not thought of beforehand and for doing exploratory research (Macdonald, 2002 ).

Defaults easily to Fisher's approach . As this approach looks superficially similar to Fisher's, it is easy to confuse both and forget what makes Neyman-Pearson's approach unique (Lehman, 1993 ). If the information provided by the alternative hypothesis—ES and β —is not taken into account for designing research with good power, data analysis defaults to Fisher's test of significance.

Null hypothesis significance testing

NHST is the most common procedure used for testing data nowadays, albeit under the false assumption of testing substantive hypotheses (Carver, 1978 ; Nickerson, 2000 ; Hubbard, 2004 ; Hager, 2013 ). NHST is, in reality, an amalgamation of Fisher's and Neyman-Pearson's theories, offered as a seamless approach to testing (Macdonald, 2002 ; Gigerenzer, 2004 ). It is not a clearly defined amalgamation either and, depending on the author describing it or on the researcher using it, it may veer more toward Fisher's approach (e.g., American Psychological Association, 2010 ; Nunnally, 1960 ; Wilkinson and the Task Force on Statistical Inference, 1999 ; Krueger, 2001 ) or toward Neyman-Pearson's approach (e.g., Cohen, 1988 ; Rosnow and Rosenthal, 1989 ; Frick, 1996 ; Schmidt, 1996 ; Cortina and Dunlap, 1997 ; Wainer, 1999 ; Nickerson, 2000 ; Kline, 2004 ).

Unfortunately, if we compare Fisher's and Neyman-Pearson's approaches vis-à-vis, we find that they are incompatible in most accounts (Table ​ (Table1). 1 ). Overall, however, most amalgamations follow Neyman-Pearson procedurally but Fisher philosophically (Spielman, 1978 ; Johnstone, 1986 ; Cortina and Dunlap, 1997 ; Hubbard, 2004 ).

Equivalence of constructs in Fisher's and Neyman-Pearson's theories, and amalgamation of constructs under NHST .

Test objectData—P(D|H )=Data—P(D|H )
NHSTData as if testing a falsifiable hypothesis—P(H |D)
ApproachA posterioriA priori
NHSTA posteriori, sometimes both
Research goalStatistical significance of research resultsDeciding between competing hypotheses
NHSTStatistical significance, also used for deciding between hypotheses
Hs under testH , to be nullified with evidenceH , to be favored against H
NHSTBoth (H = H )
Alternative hypothesisNot needed (implicitly, “No H )Needed. Provides ES and β
NHSTH posed as ‘No H ’ (ES and β sometimes considered)
Prob. distr. of testAs appropriate for H =As appropriate for H
NHSTAs appropriate for H
Cut-off pointSig identifies noteworthy results; can be gradated; can be corrected a posterioriCommon to CV , α, β, and MES; cannot be gradated; cannot be corrected a posteriori
NHSTSig = α, can be gradated, can be corrected a posteriori
Sample size calculatorNoneBased on test, ES, α, and power (1 − β)
NHSTEither
Statistic of interest -value, as evidence against H CV ( -value has no inherent meaning but can be used as a proxy instead)
NHST -value, used both as evidence against H and a proxy to accept H
Error prob.α possible, but irrelevant with single studiesα = Type I error prob. β = Type II error prob.
NHST(partly) ➥ -value = α = Type I error in single studies (β sometimes considered)
Result falls outside critical regionIgnore result as not significantAccept H if good power; conclude nothing otherwise
NHSTEither ignore result as not significant; or accept H ; or conclude nothing
Result falls in critical regionReject H Accept H (= Reject H in favor of H )
NHSTEither
Interpretation of results in critical regionEither a rare event occurred or H does not explain the research dataH explains research data better than H does (given α)
NHSTH has been proved / is true; or H has been disproved / is false; or both
Next stepsRejecting H does not automatically justify not H . Replication needed, meta-analysis is useful.Impossible to know whether α error has been made. Repeated sampling of same population needed, Monte Carlo is useful.
NHSTNone (results taken as definitive, especially if significant); further studies may be sometimes recommended (especially if results are not significant)

NHST is not only ubiquitous but very well ingrained in the minds and current practice of most researchers, journal editors and publishers (Spielman, 1978 ; Gigerenzer, 2004 ; Hubbard, 2004 ), especially in the biological sciences (Lovell, 2013 ; Ludbrook, 2013 ), social sciences (Frick, 1996 ), psychology (Nickerson, 2000 ; Gigerenzer, 2004 ) and education (Carver, 1978 , 1993 ). Indeed, most statistics textbooks for those disciplines still teach NHST rather than the two approaches of Fisher and of Neyman and Pearson as separate and rather incompatible theories (e.g., Dancey and Reidy, 2014 ). NHST has also the (false) allure of being presented as a procedure for testing substantive hypotheses (Macdonald, 2002 ; Gigerenzer, 2004 ).

In the situations in which they are most often used by researchers, and assuming the corresponding parameters are also the same, both Fisher's and Neyman-Pearson's theories work with the same statistical tools and produce the same statistical results; therefore, by extension, NHST also works with the same statistical tools and produces the same results—in practice, however, both approaches start from different starting points and lead to different outcomes (Fisher, 1955 ; Spielman, 1978 ; Berger, 2003 ). In a nutshell, the differences between Fisher's and Neyman-Pearson's theories are mostly about research philosophy and about how to interpret results (Fisher, 1955 ).

The most coherent plan of action is, of course, to follow the theory which is most appropriate for purpose, be this Fisher's or Neyman-Pearson's. It is also possible to use both for achieving different goals within the same research project (e.g., Neyman-Pearson's for tests thought of a priori, and Fisher's for exploring the data further, a posteriori), pending that those goals are not mixed up.

However, the apparent parsimony of NHST and its power to withstand threats to its predominance are also understandable. Thus, I propose two practical solutions to improve NHST: the first a compromise to improve Fisher-leaning NHST, the second a compromise to improve Neyman-Pearson-leaning NHST. A computer program such as G * Power can be used for implementing the recommendations made for both.

Improving fisher-leaning NHST

Fisher's is the closest approach to NHST; it is also the philosophy underlying common statistics packages, such as SPSS. Furthermore, because using Neyman-Pearson's concepts within NHST may be irrelevant or inelegant but hardly damaging, it requires little re-engineering. A clear improvement to NHST comes from incorporating Neyman-Pearson's constructs of effect size and of a priori sample estimation for adequate power. Estimating effect sizes (both a priori and a posteriori) ensures that researchers consider importance over mere statistical significance. A priori estimation of sample size for good power also ensures that the research has enough sensitiveness for capturing the expected effect size (Huberty, 1987 ; Macdonald, 2002 ).

Improving Neyman-Pearson-leaning NHST

NHST is particularly damaging for Neyman-Pearson's approach, simply because the latter defaults to Fisher's if important constructs are not used correctly. An importantly damaging issue is the assimilation of p -values as evidence of Type I errors and the subsequent correction of alphas to match such p -values (roving α's, Goodman, 1993 ; Hubbard, 2004 ). The best compromise for improving NHST under these circumstances is to compensate a posteriori roving alphas with a posteriori roving betas (or, if so preferred, with a posteriori roving power). Basically, if you are adjusting alpha a posteriori (roving α) to reflect both the strength of evidence (sig) and the long-run Type I error (α), you should also adjust the long-run probability of making a Type II error (roving β). Report both roving alphas and roving betas for each test, and take them into account when interpreting your research results.

NHST is very controversial, even if the controversy is not well known. A sample of helpful readings on this controversy are Christensen ( 2005 ); Hubbard ( 2004 ); Gigerenzer ( 2004 ); Goodman ( 1999 ), Louçã ( 2008 , http://www.iseg.utl.pt/departamentos/economia/wp/wp022008deuece.pdf ), Halpin and Stam ( 2006 ); Huberty ( 1993 ); Johnstone ( 1986 ), and Orlitzky ( 2012 ).

Data testing procedures represented a historical advancement for the so-called “softer” sciences, starting in biology but quickly spreading to psychology, the social sciences and education. These disciplines benefited from the principles of experimental design, the rejection of subjective probabilities and the application of statistics to small samples that Sir Ronald Fisher started popularizing in 1922 (Lehmann, 2011 ), under the umbrella of his tests of significance (e.g., Fisher, 1954 ). Two mathematical contemporaries, Jerzy Neyman and Egon Sharpe Pearson, attempted to improve Fisher's procedure and ended up developing a new theory, one for deciding between competing hypotheses (Neyman and Pearson, 1928 ), more suitable to quality control and large scale diagnostic testing (Spielman, 1973 ). Both theories had enough similarities to be easily confused (Perezgonzalez, 2014 ), especially by those less epistemologically inclined; a confusion fiercely opposed by the original authors (e.g., Fisher, 1955 )—and ever since (e.g., Nickerson, 2000 ; Lehmann, 2011 ; Hager, 2013 )—but something that irreversibly happened under the label of null hypothesis significance testing. NHST is an incompatible amalgamation of the theories of Fisher and of Neyman and Pearson (Gigerenzer, 2004 ). Curiously, it is an amalgamation that is technically reassuring despite it being, philosophically, pseudoscience. More interestingly, the numerous critiques raised against it for the past 80 years have not only failed to debunk NHST from the researcher's statistical toolbox, they have also failed to be widely known, to find their way into statistics manuals, to be edited out of journal submission requirements, and to be flagged up by peer-reviewers (e.g., Gigerenzer, 2004 ). NHST effectively negates the benefits that could be gained from Fisher's and from Neyman-Pearson's theories; it also slows scientific progress (Savage, 1957 ; Carver, 1978 , 1993 ) and may be fostering pseudoscience. The best option would be to ditch NHST altogether and revert to the theories of Fisher and of Neyman-Pearson as—and when—appropriate. For everything else, there are alternative tools, among them exploratory data analysis (Tukey, 1977 ), effect sizes (Cohen, 1988 ), confidence intervals (Neyman, 1935 ), meta-analysis (Rosenthal, 1984 ), Bayesian applications (Dienes, 2014 ) and, chiefly, honest critical thinking (Fisher, 1960 ).

Conflict of interest statement

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

  • American Psychological Association. (2010). Publication Manual of the American Psychological Association , 6th Edn Washington, DC: APA. [ Google Scholar ]
  • Bakan D. (1966). The test of significance in psychological research . Psychol. Bull . 66 , 423–437. 10.1037/h0020412 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Benjamini Y., Hochberg Y. (1995). Controlling the false discovery rate: a practical and powerful approach to multiple testing . J. R. Stat. Soc. Ser. B Stat. Methodol . 57 , 289–300. [ Google Scholar ]
  • Berger J. O. (2003). Could Fisher, Jeffreys and Neyman have agreed on testing? Stat. Sci . 18 , 1–32 10.1214/ss/1056397485 [ CrossRef ] [ Google Scholar ]
  • Carver R. P. (1978). The case against statistical significance testing . Harv. Educ. Rev . 48 , 378–399 10.1080/00220973.1993.10806591 [ CrossRef ] [ Google Scholar ]
  • Carver R. P. (1993). The case against statistical significance testing, revisited . J. Exp. Educ . 61 , 287–292 10.1080/00220973.1993.10806591 [ CrossRef ] [ Google Scholar ]
  • Christensen R. (2005). Testing Fisher, Neyman, Pearson, and Bayes . Am. Stat . 59 , 121–126 10.1198/000313005X20871 [ CrossRef ] [ Google Scholar ]
  • Cohen J. (1988). Statistical Power Analysis for the Behavioral Sciences , 2nd Edn New York, NY: Psychology Press. [ Google Scholar ]
  • Cortina J. M., Dunlap W. P. (1997). On the logic and purpose of significance testing . Psychol. Methods 2 , 161–172 10.1037/1082-989X.2.2.161 [ CrossRef ] [ Google Scholar ]
  • Cumming G. (2014). The new statistics: why and how . Psychol. Sci . 25 , 7–29. 10.1177/0956797613504966 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Dancey C. P., Reidy J. (2014). Statistics Without Maths for Psychology , 6th Edn Essex: Pearson Education. [ Google Scholar ]
  • Dienes Z. (2014). Using Bayes to get the most out of non-significant results . Front. Psychol . 5 : 781 . 10.3389/fpsyg.2014.00781 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Efron B. (1981). Nonparametric estimates of standard error: the jackknife, the bootstrap and other methods . Biometrika 68 , 589–599. 10.1093/biomet/68.3.589 [ CrossRef ] [ Google Scholar ]
  • Fisher R. A. (1932). Inverse probability and the use of likelihood . Proc. Camb. Philos. Soc . 28 , 257–261 10.1017/S0305004100010094 [ CrossRef ] [ Google Scholar ]
  • Fisher R. A. (1954). Statistical Methods for Research Workers , 12th Edn Edinburgh: Oliver and Boyd. [ Google Scholar ]
  • Fisher R. A. (1955). Statistical methods and scientific induction . J. R. Stat. Soc. Ser. B. Stat. Methodol . 17 , 69–78. [ Google Scholar ]
  • Fisher R. A. (1960). The Design of Experiments , 7th Edn Edinburgh: Oliver and Boyd. [ Google Scholar ]
  • Fisher R. A. (1973). Statistical Methods and Scientific Inference , 3rd Edn London: Collins Macmillan. [ Google Scholar ]
  • Franco A., Malhotra N., Simonovits G. (2014). Publication bias in the social sciences: unlocking the file drawer . Science 345 , 1502–1505. 10.1126/science.1255484 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Frick R. W. (1996). The appropriate use of null hypothesis testing . Psychol. Methods 1 , 379–390 10.1037/1082-989X.1.4.379 [ CrossRef ] [ Google Scholar ]
  • Gigerenzer G. (2004). Mindless statistics . J. Soc. Econ . 33 , 587–606. 10.1016/j.socec.2004.09.033 [ CrossRef ] [ Google Scholar ]
  • Gill P. M. W. (2007). Efficient calculation of p-values in linear-statistic permutation significance tests . J. Stat. Comput. Simul . 77 , 55–61. 10.1080/10629360500108053 [ CrossRef ] [ Google Scholar ]
  • Goodman S. N. (1993). P values, hypothesis tests, and likelihood: implications for epidemiology of a neglected historical debate . Am. J. Epidemiol . 137 , 485–496. [ PubMed ] [ Google Scholar ]
  • Goodman S. N. (1999). Toward evidence-based medical statistics. 1: the p-value fallacy . Ann. Intern. Med . 130 , 995–1004. 10.7326/0003-4819-130-12-199906150-00008 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Hagen R. L. (1997). In praise of the null hypothesis statistical test . Am. Psychol . 52 , 15–24 10.1037/0003-066X.52.1.15 [ CrossRef ] [ Google Scholar ]
  • Hager W. (2013). The statistical theories of Fisher and of Neyman and Pearson: a methodological perspective . Theor. Psychol . 23 , 251–270 10.1177/0959354312465483 [ CrossRef ] [ Google Scholar ]
  • Halpin P. F., Stam H. J. (2006). Inductive inference or inductive behavior: fisher and Neyman–Pearson approaches to statistical testing in psychological research (1940-1960) . Am. J. Psychol . 119 , 625–653. 10.2307/20445367 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Holland B. S., Copenhaver M. D. (1987). An improved sequentially rejective Bonferroni test procedure . Biometrics 43 , 417–423 10.2307/2531823 [ CrossRef ] [ Google Scholar ]
  • Hubbard R. (2004). Alphabet soup. Blurring the distinctions between p's and α's in psychological research . Theor. Psychol . 14 , 295–327 10.1177/0959354304043638 [ CrossRef ] [ Google Scholar ]
  • Huberty C. J. (1987). On statistical testing . Educ. Res . 8 , 4–9 10.3102/0013189X016008004 [ CrossRef ] [ Google Scholar ]
  • Huberty C. J. (1993). Historical origins of statistical testing practices: the treatment of Fisher versus Neyman-Pearson views in textbooks . J. Exp. Educ . 61 , 317–333 10.1080/00220973.1993.10806593 [ CrossRef ] [ Google Scholar ]
  • Johnstone D. J. (1986). Tests of significance in theory and practice . Statistician 35 , 491–504 10.2307/2987965 [ CrossRef ] [ Google Scholar ]
  • Johnstone D. J. (1987). Tests of significance following R. A. Fisher . Br. J. Philos. Sci . 38 , 481–499 10.1093/bjps/38.4.481 [ CrossRef ] [ Google Scholar ]
  • Jones L. V., Tukey J. W. (2000). A sensible formulation of the significance test . Psychol. Methods 5 , 411–414. 10.1037/1082-989X.5.4.411 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Kline R. B. (2004). Beyond Significance Testing. Reforming Data Analysis Methods in Behavioral Research . Washington, DC: APA. [ Google Scholar ]
  • Krueger J. (2001). Null hypothesis significance testing. On the survival of a flawed method . Am. Psychol . 56 , 16–26. 10.1037/0003-066X.56.1.16 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Kruskal W. (1980). The significance of Fisher: a review of R. A. Fisher: the life of a scientist . J. Am. Stat. Assoc . 75 , 1019–1030 10.2307/2287199 [ CrossRef ] [ Google Scholar ]
  • Lehman E. L. (1993). The Fisher, Neyman-Pearson theories of testing hypothesis: one theory or two? J. Am. Stat. Assoc . 88 , 1242–1249 10.1080/01621459.1993.10476404 [ CrossRef ] [ Google Scholar ]
  • Lehmann E. L. (2011). Fisher, Neyman, and the creation of classical statistics . New York, NY: Springer. [ Google Scholar ]
  • Lindley D. V. (1965). Introduction to Probability and Statistics from a Bayesian Viewpoint, Part 1: Probability; Part 2: Inference . Cambridge: Cambridge University Press. [ Google Scholar ]
  • Lindquist E. F. (1940). Statistical Analysis in Educational Research . Boston, MA: Houghton Mifflin. [ Google Scholar ]
  • Louçã F. (2008). Should the Widest Cleft in Statistics - How and Why Fisher Opposed Neyman and Pearson . WP/02/2008/DE/UECE, Technical University of Lisbon. Available online at: http://www.iseg.utl.pt/departamentos/economia/wp/wp022008deuece.pdf
  • Lovell D. P. (2013). Biological importance and statistical significance . J. Agric. Food Chem . 61 , 8340–8348. 10.1021/jf401124y [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Ludbrook J. (2013). Should we use one-sided or two-sided p values in tests of significance? Clin. Exp. Pharmacol. Physiol . 40 , 357–361 10.1111/1440-1681.12086 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Macdonald R. R. (1997). On statistical testing in psychology . Br. J. Psychol . 88 , 333–347 10.1111/j.2044-8295.1997.tb02638.x [ CrossRef ] [ Google Scholar ]
  • Macdonald R. R. (2002). The incompleteness of probability models and the resultant implications for theories of statistical inference . Underst. Stat . 1 , 167–189 10.1207/S15328031US0103_03 [ CrossRef ] [ Google Scholar ]
  • Newman I., Fraas J., Herbert A. (2001). Testing non-nil null hypotheses with T tests of group means: a monte carlo study , in Annual Meeting Mid-Western Educational Research Association (Chicago, IL).
  • Neyman J. (1935). On the problem of confidence intervals . Ann. Math. Stat . 6 , 111–116 10.1214/aoms/1177732585 [ CrossRef ] [ Google Scholar ]
  • Neyman J. (1942). Basic ideas and some recent results of the theory of testing statistical hypotheses . J. R. Stat. Soc . 105 , 292–327 10.2307/2980436 [ CrossRef ] [ Google Scholar ]
  • Neyman J. (1953). First Course in Probability and Statistics . New York, NY: Henry Holt. [ Google Scholar ]
  • Neyman J. (1955). The problem of inductive inference . Commun. Pure Appl.Math . III , 13–46 10.1002/cpa.3160080103 [ CrossRef ] [ Google Scholar ]
  • Neyman J. (1956). Note on an article by Sir Ronald Fisher . J. R. Stat. Soc. Ser. B. Stat. Methodol . 18 , 288–294. [ Google Scholar ]
  • Neyman J. (1967). R. A. Fisher (1890-1962): an appreciation . Science 156 , 1456–1460. 10.1126/science.156.3781.1456 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Neyman J., Pearson E. S. (1928). On the use and interpretation of certain test criteria for purposes of statistical inference: part I . Biometrika 20A , 175–240 10.2307/2331945 [ CrossRef ] [ Google Scholar ]
  • Neyman J., Pearson E. S. (1933). On the problem of the most efficient tests of statistical hypotheses . Philos. Trans. R. Soc. Lond. A 231 , 289–337 10.1098/rsta.1933.0009 [ CrossRef ] [ Google Scholar ]
  • Nickerson R. S. (2000). Null hypothesis significance testing: a review of an old and continuing controversy . Psychol. Methods 5 , 241–301. 10.1037/1082-989X.5.2.241 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Nunnally J. (1960). The place of statistics in psychology . Educ. Psychol. Meas . 20 , 641–650 10.1177/001316446002000401 [ CrossRef ] [ Google Scholar ]
  • Orlitzky M. (2012). How can significance tests be deinstitutionalized? Organ. Res. Methods 15 , 199–228 10.1177/1094428111428356 [ CrossRef ] [ Google Scholar ]
  • Pearson E. S. (1955). Statistical concepts in the relation to reality . J. R. Stat. Soc. Ser. B Stat. Methodol . 17 , 204–207. [ Google Scholar ]
  • Pearson E. S. (1990). ‘Student’ a Statistical Biography of William Sealy Gosset . Oxford: Claredon Press. [ Google Scholar ]
  • Perezgonzalez J. D. (2014). A reconceptualization of significance testing . Theor. Psychol . 24 , 852–859 10.1177/0959354314546157 [ CrossRef ] [ Google Scholar ]
  • Perezgonzalez J. D. (2015). Confidence intervals and tests are two sides of the same research question . Front. Psychol . 6 : 34 . 10.3389/fpsyg.2015.00034 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Rosenthal R. (1984). Meta-Analytic Procedures for Social Research . Beverly Hills, CA: Sage. [ Google Scholar ]
  • Rosnow R. L., Rosenthal R. (1989). Statistical procedures and the justification of knowledge in psychological science . Am. Psychol . 44 , 1276–1284 10.1037/0003-066X.44.10.1276 [ CrossRef ] [ Google Scholar ]
  • Savage I. R. (1957). Nonparametric statistics . J. Am. Stat. Assoc . 52 , 331–344 10.1080/01621459.1957.10501392 [ CrossRef ] [ Google Scholar ]
  • Schmidt F. L. (1996). Statistical significance testing and cumulative knowledge in psychology: implications for training of researchers . Psychol. Methods 1 , 115–129 10.1037/1082-989X.1.2.115 [ CrossRef ] [ Google Scholar ]
  • Spielman S. (1973). A refutation of the Neyman-Pearson theory of testing . Br. J. Philos. Sci . 24 , 201–222 10.1093/bjps/24.3.201 [ CrossRef ] [ Google Scholar ]
  • Spielman S. (1978). Statistical dogma and the logic of significance testing . Philos. Sci . 45 , 120–135 10.1086/288784 [ CrossRef ] [ Google Scholar ]
  • Tukey J. W. (1977). Exploratory Data Analysis . Reading, MA: Addison-Wesley. [ Google Scholar ]
  • Wainer H. (1999). One cheer for null hypothesis significance testing . Psychol. Methods 4 , 212–213 10.1037/1082-989X.4.2.212 [ CrossRef ] [ Google Scholar ]
  • Wald A. (1950). Statistical Decision Functions . Ney York, NY: Wiley. [ Google Scholar ]
  • Wilkinson L., the Task Force on Statistical Inference. (1999). Statistical methods in psychology journals. Guidelines and explanations . Am. Psychol . 54 , 594–604 10.1037/0003-066X.54.8.594 [ CrossRef ] [ Google Scholar ]
  • Williams R. H., Zimmerman D. W., Ross D. C., Zumbo B. D. (2006). Twelve British Statisticians . Raleigh, NC: Boson Books. [ Google Scholar ]

Tools, Technologies and Training for Healthcare Laboratories

Westgard

  • Z-Stats / Basic Statistics

Z-8: Two-Sample and Directional Hypothesis Testing

This lesson describes some refinements to the hypothesis testing approach that was introduced in the previous lesson. The truth of the matter is that the previous lesson was somewhat oversimplified in order to focus on the concept and general steps in the hypothesis testing procedure. With that background, we can now get into some of the finer points of hypothesis testing.

EdD, Assistant Professor Clinical Laboratory Science Program, University of Louisville Louisville, Kentucky October 1999

Refinements in calculations, refinements in procedure, a two-tailed test for our mice, a one-tailed test for our mice.

  • Critical t-values for one-tailed and two-tailed tests
  • About the Author

Two sample case

The "two sample case" is a special case in which the difference between two sample means is examined. This sounds like what we did in the last lesson, but we actually looked at the difference between an observed or sample group mean and a control group mean, which was treated as if it were a population mean (rather than an observed or sample mean). There are some fine points that need to be considered about the calculations and the procedure.

Mathematically, there are some differences in the formula that should be used. Recall the form of the equation for calculating a t-value:

where Xbar is the mean of the experimental group, and µ is a population parameter or the mean of the population. In the last lesson, we substituted the control group mean XbarB for µ. However, the control group was actually a sample from the population of all such mice (in this example) and following suit, the experimental group was also just a sample from its population. Phrasing the situation this way, there are now two sets of differences that must be considered: the difference between sample means or XbarA-XbarB and the difference between population means from which these samples came or µA-µB. Those expressions should be substituted into the t calc formula to give the proper mathematical form:

  • In this new equation, the difference between means (Xbar A - XbarB) is called bias and is important in determining test accuracy, so even though this discussion is getting more complicated, the statistic that we are deriving is very important.
  • When the square root is taken, the result is called the standard deviation of the difference between means , SDd , which is shown by the formula:
  • The proper equation for calculating the t-value for the two-sample case then becomes:

Remember the steps for testing a hypothesis are: (1) State the hypotheses; (2) Set the criterion for rejection of Ho; (3) Compute the test statistic; (4) Decide about Ho.

  • The null hypothesis can be stated as: Ho: µA = µB or µA - µB = 0. But it may be more revealing to say Ho: (XbarA-XbarB) - (µA - µB) = 0. The difference between the sample means minus the difference between the population means equals zero.
  • The alternative hypothesis can be stated as: Ha: µA is not equal to µB or Ha: (XbarA-XbarB) - (µA - µB) is not equal to 0, i.e., the means of the two groups are not equal.
  • The criteria for rejection or the alpha level is 0.05.
  • The test statistic is computed as:
  • Since it is hypothesized that the two methods are comparable and the difference between the means of the two populations is zero (µA - µB = 0), the calculation can be simplified as follows:
  • If this t calc is greater than 1.96 then the null hypothesis of no difference can be overturned (p0.05).
  • Even though we have used several different mathematical formulae, the interpretations are the same as before.

Directional hypothesis testing vs non-directional testing

Remember the example of testing the effect of antibiotics on mice in Lesson 7. The point of the study was to find out if the mice who were treated with the antibiotic would outlive those who were not treated (i.e., the control group). Are you surprised that the researcher did not hypothesize that the control group might outlive the treatment group? Would it make any difference in how the hypothesis testing were carried out? These questions raise the issue of directional testing, or one-tailed vs two-tailed tests.

The issue of two-side vs one-side tests becomes important when selecting the critical t-value. In the earlier discussion of this example, the alpha level was set to 0.05, but that 0.05 was actually divided equally between the left and right tails of the distribution curve. The condition being tested is that group A has a "different" life span as compared to group B, which represents a two-tailed test as illustrated in Figure 8-2.

Critical t-values for one-tailed and two tailed tests

This can be confusing and it should be helpful to think about having one or two gates on the curve. For example, for an alpha level of 0.05 and a two-tail test, there are two gates - one at -1.96 and one at +1.96. If you "walk" out of either of those gates, then you have demonstrated significance at p=0.05 or less. For a one-tail test, there is only one gate at +1.65. If you "walk" out of this gate, you have demonstrated significance at p=0.05 or less.

To reiterate, if you are standing right at the gate (1.96) for a two-tail test, then you have just barely met the p=0.05 requirement. However, if you are standing at the 1.96 point when running a one-tail test, then you have already exceeded the 1.65 gate and the probability must be even more significant, say p=0.025. It's important to find the critical t-value that is correct for the intended directional nature of the test.

About the author: Madelon F. Zady

Madelon F. Zady is an Assistant Professor at the University of Louisville, School of Allied Health Sciences Clinical Laboratory Science program and has over 30 years experience in teaching. She holds BS, MAT and EdD degrees from the University of Louisville, has taken other advanced course work from the School of Medicine and School of Education, and also advanced courses in statistics. She is a registered MT(ASCP) and a credentialed CLS(NCA) and has worked part-time as a bench technologist for 14 years. She is a member of the: American Society for Clinical Laboratory Science, Kentucky State Society for Clinical Laboratory Science, American Educational Research Association, and the National Science Teachers Association. Her teaching areas are clinical chemistry and statistics. Her research areas are metacognition and learning theory.

  • "Westgard Rules"
  • Basic QC Practices
  • High Reliability
  • "Housekeeping"
  • Maryland General
  • Method Validation
  • Quality Requirements and Standards
  • Quality of Laboratory Testing
  • Guest Essay
  • Risk Management
  • Sigma Metric Analysis
  • Quality Standards
  • Basic Planning for Quality
  • Basic Method Validation
  • Quality Management
  • Advanced Quality Management / Six Sigma
  • Patient Safety Concepts
  • Quality Requirements
  • CLIA Final Rule
  • Course Orders
  • Feedback Form

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • View all journals
  • Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • Open access
  • Published: 07 September 2024

Loss of DOCK2 potentiates Inflammatory Bowel Disease–associated colorectal cancer via immune dysfunction and IFNγ induction of IDO1 expression

  • Antonia M. D. Churchhouse 1 ,
  • Caroline V. Billard 1 ,
  • Toshiyasu Suzuki 2 , 3 ,
  • Sebastian Ö. G. Pohl   ORCID: orcid.org/0000-0002-4294-9498 1 ,
  • Nora J. Doleschall 1 ,
  • Kevin Donnelly 1 ,
  • Colin Nixon   ORCID: orcid.org/0000-0002-8085-2160 2 ,
  • Mark J. Arends   ORCID: orcid.org/0000-0002-6826-8770 1 ,
  • Shahida Din   ORCID: orcid.org/0000-0003-2855-3400 4 ,
  • Kathryn Kirkwood 5 ,
  • Jair Marques Junior   ORCID: orcid.org/0000-0002-8095-8475 1 ,
  • Alex Von Kriegsheim   ORCID: orcid.org/0000-0002-4952-8573 1 ,
  • Seth B. Coffelt   ORCID: orcid.org/0000-0003-2257-2862 2 , 3 &
  • Kevin B. Myant   ORCID: orcid.org/0000-0001-8017-1093 1  

Oncogene ( 2024 ) Cite this article

11 Altmetric

Metrics details

  • Colorectal cancer
  • Inflammation

Inflammatory Bowel Disease-associated colorectal cancer (IBD-CRC) is a known and serious complication of Inflammatory Bowel Disease (IBD) affecting the colon. However, relatively little is known about the pathogenesis of IBD-associated colorectal cancer in comparison with its sporadic cancer counterpart. Here, we investigated the function of Dock2 , a gene mutated in ~10% of IBD-associated colorectal cancers that encodes a guanine nucleotide exchange factor (GEF). Using a genetically engineered mouse model of IBD-CRC, we found that whole body loss of Dock2 increases tumourigenesis via immune dysregulation. Dock2 -deficient tumours displayed increased levels of IFNγ-associated genes, including the tryptophan metabolising, immune modulatory enzyme, IDO1, when compared to Dock2 -proficient tumours. This phenotype was driven by increased IFNγ-production in T cell populations, which infiltrated Dock2 -deficient tumours, promoting IDO1 expression in tumour epithelial cells. We show that IDO1 inhibition delays tumourigenesis in Dock2 knockout mice, and we confirm that this pathway is conserved across species as IDO1 expression is elevated in human IBD-CRC and in sporadic CRC cases with mutated DOCK2 . Together, these data demonstrate a previously unidentified tumour suppressive role of DOCK2 that limits IFNγ-induced IDO1 expression and cancer progression, opening potential new avenues for therapeutic intervention.

Similar content being viewed by others

non directional two tailed hypothesis

The role of the Hippo pathway in the pathogenesis of inflammatory bowel disease

non directional two tailed hypothesis

Inactivation of Interferon Regulatory Factor 1 Causes Susceptibility to Colitis-Associated Colorectal Cancer

non directional two tailed hypothesis

IL-36 signalling enhances a pro-tumorigenic phenotype in colon cancer cells with cancer cell growth restricted by administration of the IL-36R antagonist

Introduction.

Inflammatory Bowel Disease–associated colorectal cancer (IBD-CRC) is a long term colonic complication of inflammatory Bowel Disease (IBD) [ 1 ]. IBD is a complex, chronic disease affecting the gastrointestinal tract, characterised by immune dysregulation as a result of genetic, environmental and microbiotal factors [ 2 ], and is generally classified into two subtypes: Ulcerative Colitis, and Crohn’s Disease. IBD-CRC occurs on a background of colonic inflammation, and occurs in younger patients [ 3 ], and is associated with increased mortality compared to sporadic colorectal cancer [ 3 , 4 ], yet its pathogenesis is poorly understood. There is therefore an urgent need to understand IBD-CRC in greater detail.

The mutational spectrum of IBD-CRC is different from sporadic colorectal cancer with alterations in TP53 being the most common driver event and APC and KRAS being mutated at significantly lower rates [ 5 , 6 , 7 ]. In addition, alterations in the RHO/RAC signalling pathway are frequently associated with IBD-CRC with exome sequencing studies suggesting a prevalence of 50–70% of tumours having alterations in this pathway [ 6 , 7 ]. It should be noted however that these data do not imply significant mutation, and the functional effects of these mutations are unknown. RAC proteins have a critical role in several cellular processes including migration, apoptosis, proliferation, and invasion [ 8 , 9 , 10 , 11 ] but their role in the etiology of IBD-CRC is poorly understood.

Of particular interest are mutations in the dedicator of cytokinesis 2 ( DOCK2 ) gene, which have been found in ~10% of IBD-CRC cases [ 6 , 7 ]. DOCK2 is a 212 kDa protein expressed in both human and mouse, found mainly in peripheral blood cells, and to a lesser extent, the thymus and spleen [ 12 ]. DOCK proteins act primarily as guanine nucleotide exchange factors (GEFs) for the RAC/RHO GTPases [ 13 ]. GEF activity involves catalysing the release of GDP in exchange for GTP on RAC, resulting in RAC activation [ 11 ]. DOCK2 specifically is involved in both the innate and adaptive immune responses and plays a critical role in activation and proliferation of T lymphocytes [ 14 ]. DOCK2 was recently identified as a key driver gene in human IBD using network predictive modelling [ 15 ]. Dock2 —/— mice are also more sensitive to colitis as a result of Citrobacter rodentium infection [ 16 ]. As DOCK2 has been linked to both IBD and IBD-CRC in the context of its known physiological effect on immune cells, it is a particularly attractive target to study, and to our knowledge, the functional role of DOCK2 on the development of IBD-CRC has not been previously examined.

Here, we demonstrate loss of Dock2 in a Dextran Sodium Sulphate (DSS) colitis-induced mouse model of IBD-CRC increases tumour formation. We find this increased tumourigenesis is associated with CD3 + T cell infiltration, increased interferon gamma signalling, elevated expression of the IFNγ target IDO1 and increased tryptophan metabolism in Dock2 -deficient tumours. In addition, we show elevated production of IFNγ in immune cell populations, including γδ and CD8 T cells, under normal homeostatic conditions suggesting loss of Dock2 leads to immune cell dysfunction and aberrant IFNγ signalling. IDO1 is elevated in human IBD-CRC patients and pharmacological inhibition of IDO1 activity in Dock2- deficient mice abrogates tumourigenesis. Together, these results outline a novel functional role for aberrant IFNγ producing immune dysfunction in promoting IBD-CRC via modulation of IDO1-mediated tryptophan metabolism and suggest potential therapeutic avenues for targeting human disease.

Loss of Dock2 leads to increased tumour formation in vivo

To determine whether loss of Dock2 promotes inflammation-associated CRC development we crossed mice carrying a whole mouse ‘deletion first’ Dock2 tm1a allele to a well characterised villin-cre ERT2 Apc fl/+ intestinal CRC model [ 17 ] (Fig. S1A , B ) generating cohorts of control villin-cre ERT2 Apc fl/+ ( Vil Apc ) and experimental villin-cre ERT2 Apc fl/+ Dock2 tm1a/tm1a ( Vil Apc Dock2 ) mice. Following induction with tamoxifen, mice were subjected to two rounds of colitis-inducing DSS treatment, aged to 47 days where after colons were harvested for histological analysis (Fig. 1A ). Loss of Dock2 expression due to the ‘deletion first’ allele was confirmed in colonic tissue by mRNA and protein analysis and in thymus by RT-qPCR (Fig. S1C – E ). Vil Apc Dock2 mice developed more colonic tumours than controls and had an increased overall tumour burden (Fig. 1B–D ). Average tumour size and tumour cell proliferation was not different between groups (Fig. 1E–G ) suggesting increased tumourigenesis was not due to accelerated tumour growth (Fig. 1F, G ).

figure 1

A Schematic detailing the Apc loss-mediated mouse model of IBD-associated colorectal cancer, including inducible gene modification with tamoxifen, two rounds of 0.5% DSS, and termination at day 47 post DSS administration. B Representative H&E-stained colonic Swiss rolls, detailing tumour burden in Vil Apc and Vil Apc Dock2 mice (black arrows). Scale bars are 2.5 mm. C Tumour number in Vil Apc and Vil Apc Dock2 mice. N = 11 vs 11 mice. D Total tumour burden in Vil Apc and Vil Apc Dock2 mice. N = 11 vs 11 mice. E Representative BrdU-stained Vil Apc and Vil Apc Dock2 tumours. Scale bars are 50 μm. F Quantification of BrdU staining. N = 7 vs 11 mice. G Average tumour size in Vil Apc and Vil Apc Dock2 mice. N = 11 vs 11 mice. Data represented as mean and error bars SD. All statistical analysis for this figure was performed using two-tailed Mann–Whitney test. Exact p values are indicated in the panels.

Tumours in mice lacking Dock2 show an immune phenotype, elevated interferon gamma signalling, and upregulation of IDO1

To identify potential molecular mechanisms mediating increased tumourigenesis following Dock2 deletion we carried out RNAseq on 5 vs 9 colonic tumours dissected from Vil Apc and Vil Apc Dock2 mice. This analysis revealed 80 transcripts with altered expression (FC > ±1.5, padj < 0.05; 65 upregulated, 15 downregulated) (Fig. 2A and Table S1 ). Dock2 was the most downregulated gene, with a logFC of −2.89 and an adjusted p value = 8.87 × 10 −35 , in line with the minimal Dock2 expression observed in the Vil Apc Dock2 tumours (Fig. S1C , D ). Gene ontology (GO) analysis of our dataset identified enrichment of numerous immune related pathways as being activated in Dock2 deficient tumours (Fig. 2B and Table S2 ). These included pathways involved in immune system processes, MHC protein binding and response to interferon signalling. Additionally, transcription factor binding site (TFBS) analysis demonstrated an enrichment of putative targets of interferon regulatory factors (IRFs) (Fig. 2B and Table S2 ). Furthermore, gene set enrichment analysis (GSEA) identified enrichment of interferon gamma and alpha response hallmark gene sets with genes upregulated in our dataset (Fig. S2A ; IFNγ p < 0.001 and IFNα p = 0.005). Together, this suggests loss of Dock2 leads to dysregulation of immune system processes and, in particular, activation of interferon signalling. Quantitative RT-PCR was used to validate a selection of the identified interferon related genes with all tested genes showing the expected transcriptional changes (Fig. 2C ). One of the most upregulated genes following Dock2 loss was indoleamine 2, 3-dioxygenase 1 ( Ido1 ). Ido1 is a previously described target of IFNγ signalling with known tumour promoting, immunomodulatory functions [ 18 , 19 ]. Therefore, we chose to investigate IDO1 in more detail. We further validated this upregulation at protein level observing upregulation of IDO1 protein in Dock2 deficient tumours by IHC and Western blot (Fig. 2D, E and Fig. S2B , C ). To define in better detail the cellular compartment expressing IDO1 we co-stained Vil Apc Dock2 tumour tissue for epithelial and immune cell markers alongside IDO1. IDO1 expression was exclusively co-localised with CDH1 expression and we did not observe co-expression with CD3 or CD8 markers (Fig. S2D , E ). Together, this suggests an induction of IFNγ signalling in the tumour epithelium of Dock2 deficient mice. We also observed increased Ido1 but not Ifng expression in normal tissue adjacent to tumours demonstrating this effect is not restricted to tumour cells (Fig. S2F ).

figure 2

A Heatmap of genes with significantly altered expression in Vil Apc Dock2 tumour identified by RNAseq. B Gene ontology analysis of Vil Apc vs Vil Apc Dock2 tumours. The top 5 non-redundant Biological Processes (BP), Molecular Functions (MF), Cellular Components (CC) and Putative Transcription Factor Binding Sites (TFBS) are listed and corresponding –log10 padj shown. C RT-qPCR analysis of IFNγ target gene expression in Vil Apc and Vil Apc Dock2 tumours. N = 5 vs 6 tumours ( Ifng ) or 5 vs 9 tumours (others). D Representative IDO1-stained Vil Apc and Vil Apc Dock2 tumours. Scale bars are 500 μm and 100 μm (zoom). E Quantification of IDO1 staining. N = 6 vs 9 tumours. Data represented as mean and error bars SD. All statistical analysis for this figure was performed using two-tailed Mann–Whitney test. Exact p values are indicated in the panels.

To further investigate this, we determined whether colonic tumour epithelial cells respond to exogenous IFNγ. Colonic organoids were derived from control and Dock2 tm1a/tm1a ( Dock2 ) mice and Apc deleted using CRISPR to mimic tumour development in an inflammation naïve setting (Fig. S3A ). Apc KO and Apc KO Dock2 organoids were treated with mouse IFNγ for 24 h. There were no changes in organoid growth (Fig. 3A, B ) but we observed an upregulation of genes identified in our RNAseq dataset confirming they are IFNγ targets in colonic epithelial cells and indicating a key role for type II interferon signalling in the epithelial response to interferon stimulation (Fig. 3C ). This included increased Ido1 expression which was also observed at protein level (Fig. 3D–H ). Interestingly, whilst there was a modest increase in the activation of some IFNγ target genes in IFNγ treated Apc KO Dock2 organoids compared to IFNγ treated controls this was not the case for Ido1 mRNA or protein, the activation of which was broadly the same (Fig. 3C–E ). This was in contrast to our in vivo data showing an upregulation of Ido1 in Dock2 deficient mice. Thus, the responsiveness of Dock2 deficient and normal tumour organoids to IFNγ stimulation is broadly similar. Together, this suggests the increased IFNγ response observed in vivo may be due to elevated infiltration of IFNγ producing cells in Dock2 deficient tumours. To further test this, we analysed the dose dependency of Apc KO organoids to IFNγ stimulation. We found that IDO1 protein and mRNA expression was induced in a dose dependent manner by increasing concentrations of IFNγ (Fig. S3B – D ). Additionally, we found that Apc KO organoids did not express detectable levels of Ifng mRNA, even following inflammatory stimulus with TNFα. Together, this supports the hypothesis that increased IFNγ production, from non-cell autonomous sources, is driving elevated IDO1 expression in tumour epithelial cells in Dock2 deficient mice. Therefore, we sought to identify potential sources of this IFNγ production.

figure 3

A Collage overview of Apc KO and Apc KO Dock2 organoids. B Quantification of Apc KO and Apc KO Dock2 organoid size change 24 h post IFNγ treatment. N = 30 vs 30 vs 30 vs 30 organoids. C RT-qPCR analysis of IFNγ target gene expression in Apc KO and Apc KO Dock2 organoids 24 h post IFNγ treatment. N = 3 vs 3 vs 3 vs 3 independent technical replicates and 3 vs 3 vs 2 vs 3 independent technical replicates ( Tap1 ). D Western blot analysis of IDO1 expression in Apc KO and Apc KO Dock2 organoids 24 h post IFNγ treatment. E Quantification of IDO1 Western blot. N = 3 vs 3 vs 3 vs 3 independent technical replicates. F Immunofluorescence staining of Ido1 expression in Apc KO organoids 24 h post IFNγ treatment. G Quantification of Ido1 staining intensity. N = 572 vs 232 organoid cells. H Quantification of % Ido1 positive cells. N = 3 vs 3 independent technical replicates. Data represented as mean and error bars SD. Statistical analysis for ( B ), ( C ) and ( E ) was performed using ordinary one-way ANOVA with Tukey’s multiple comparisons. Statistical analysis for ( G ) and ( H ) was performed using student T-test *p < 0.05, **p < 0.01, ***p < 0.001, ****p < 0.0001.

Tumours in mice lacking Dock2 are infiltrated by CD3+ and γδ T lymphocytes

As DOCK2 primarily functions as an immune modulator we examined whether immune cell dysregulation might be responsible for the observed IFNγ stimulation and thus impacting tumour formation in Dock2 deleted mice. We first investigated lymphoid and myeloid cell populations using immunohistochemistry and found an increase in intratumoural CD3 + T cell infiltration in Vil Apc Dock2 tumours (Fig. 4A, B ) but not of macrophages or neutrophils (Fig. S4A , B ). Further investigations of different T cell subtypes demonstrated that whilst CD4+ and CD8 + T cell infiltration was the same between groups (Fig. S4C , D ), there were more γδ T cells in Vil Apc Dock2 tumours than control tumours (Fig. 4C, D ). The expression of Trgv1 , the gene encoding T cell receptor Vγ1 chain, but not Trgv4, Trgv5, Trgv6 or Trgv7 was also upregulated in tumours of Vil Apc Dock2 mice (Fig. 4E and S4E ). Infiltrating Vγ1 γδ T cells have previously been identified as a potential source of IFNγ in tumours suggesting this immune cell population may be partly mediating IFNγ signalling in Dock2 deficient tumours.

figure 4

A Representative CD3-stained Vil Apc and Vil Apc Dock2 tumours. Scale bars are 50 μm ( B ) Quantification of CD3 staining. Values represent % of total tumour cells. N = 6 vs 10 mice. C Representative images of RNAscope for TRDC in tumours of Vil Apc and Vil Apc Dock2 mice. Scale bars are 100 μm. D Quantification of TRDC staining. Values represent % of total tumour cells. N = 7 vs 20 tumours. E RT-qPCR analysis of Trgv1 expression in Vil Apc and Vil Apc Dock2 tumours. N = 5 vs 9 tumours. Data represented as mean and error bars SD. All statistical analysis for this figure was performed using two-tailed Mann–Whitney test. Exact p values are indicated in the panels.

γδ T cells from mice lacking Dock2 produce more IFNγ in vivo

We next examined in more detail whether γδ T-cells are a source of IFNγ in our Dock2- deficient mouse model and whether this phenotype is intrinsic to Dock2 loss or requires exogenous inflammatory stimulus. To address this, we analysed the effects of Dock2 deletion on colonic immune cell populations under both homeostatic and inflammatory conditions. Acute colonic damage and inflammation was induced by treatment with 2% DSS. Monitoring of mice during treatment indicated no significant differences in disease severity (as indicated by weight loss) between control and Dock2 deficient mice (Fig. 5A ). After 7 days treatment (5 days DSS + 2 days normal drinking water) mice were culled and colons harvested (Fig. S5A ). Histological analysis indicated no significant difference in the extent of mucosal damage or CD3 + T cell infiltration between control and Dock2 deficient mice (Figs. 5B, C and S5 B, C ) demonstrating loss of Dock2 does not increase susceptibility to DSS induced colitis. To investigate the potential for immune cell dysregulation during acute colitis in more detail we carried out flow cytometry analysis of different lymphoid populations (Fig. S5D ). Consistent with our previous findings in cancer models, we found γδ T cells were more abundant in Dock2 deficient than control colons (Fig. 5D, E ). Notably, this increase was found under normal, homoeostatic conditions in the absence of DSS induced inflammatory stimulus. Indeed, upon DSS treatment, the proportion of γδ T cells was not different between control and Dock2 deficient mice (Fig. 5D, E ). We next determined whether γδ T cells could act as a source of IFNγ explaining the increased IFNγ activation observed in Dock2 deficient mice. Flow cytometry analysis showed that a higher proportion of both γδ T cells and CD8 + T cells expressed IFNγ in Dock2 deficient colons. Again, this increase was present in the absence of DSS treatment, suggesting Dock2 loss leads to an intrinsic immune cell dysfunction manifested by increased infiltration of IFNγ producing γδ T cells (Fig. 5F–I ). To determine whether other immune populations could contribute to increased IFNγ production we further analysed IFNγ levels in CD3 positive and CD3 negative cells by flow cytometry. We found that the majority of IFNγ producing cells were CD3 positive but IFNγ was also produced by CD3 negative cells and the proportion of these cells was also significantly higher in Dock2 deficient mice (Fig. S5F , G ). Together, this suggests Dock loss leads to immune dysregulation, characterised by increased IFNγ production in multiple immune subtypes including, but not restricted to, γδ T cells and CD8 + T cells.

figure 5

A Comparative weights of WT and Dock2 mice during treatment with 2% DSS. B Representative H&E-stained colonic Swiss rolls, detailing extent of colitis in WT and Dock2 mice. Representative areas of epithelial erosion are highlighted. Scale bars are 2.5 mm and 100 μm (zoom). C Quantification of colitic area in WT and Dock2 mice. N = 17 vs 14 mice. D Representative plots of CD3+ γδ T cells in WT and Dock2 colons at baseline or following acute DSS treatment. E Quantification of CD3+ γδ T cells. N = 3 vs 3 vs 3 vs 3 mice. F Representative plots of IFNγ producing γδ T cells in WT and Dock2 colons at baseline or following acute DSS treatment. G Quantification of IFNγ producing γδ T cells. N = 3 vs 3 vs 3 vs 3 mice. H Representative plots of IFNγ producing CD8 + T cells in WT and Dock2 colons at baseline or following acute DSS treatment. I Quantification of IFNγ producing CD8 + T cells. N = 3 vs 3 vs 3 vs 3 mice. J RT-qPCR analysis of IFNγ target gene expression in WT and Dock2 colons at baseline or following acute DSS treatment. N = 6 vs 6 vs 17 vs 14 mice. Data represented as mean and error bars SD. Statistical analysis for ( C ) was performed using two-tailed Mann–Whitney test. Statistical analysis for ( E ), ( G ), ( I ) and ( J ) was performed using ordinary one-way ANOVA with Tukey’s multiple comparisons. Exact p values are indicated in the panels.

To determine whether this led to increased IFNγ target gene expression in the colon we carried out transcriptional analysis of previously identified IFNγ responsive genes. Under homeostatic conditions Dock2 loss was not sufficient to increase IFNγ signalling in the colon but following DSS treatment a number of IFNγ responsive genes were induced (Fig. 5J ). Therefore, Dock2 deficient colonic epithelium is not inherently IFNγ activated, rather, it responds to inflammation to induce and maintain this state. To investigate more broadly, we also examined expression of γδ T cell markers, and some IFNγ responsive genes in thymus from untreated control and Dock2 deficient mice. These mice exhibit significantly higher expression of various γδ T cell markers but not IFNγ target genes (Fig. S5H ). Together, these data suggest Dock2 deficient mice have increased infiltration of IFNγ-producing immune cells, but exogenous inflammatory stimuli are needed to induce robust IFNγ activation and IDO1 expression in the colonic epithelium.

IDO1 expression is increased in human IBD-CRC and IDO1 inhibition abrogates tumourigenesis in Dock2 deficient mice

To determine whether the observations from our model are recapitulated in human disease we investigated the expression of the IFNγ target IDO1 in a set of human samples encompassing different stages of IBD-CRC disease progression. IHC analysis showed that IDO1 expression is significantly increased in inflamed tissue compared to normal colon and this increased expression persists through to the development of cancer (Fig. 6A, B ).

figure 6

A Representative IDO1-stained human normal colon, inflamed colon, dysplastic colon and IBD associated cancer (all regions of the same human resection specimen). Scale bars are 250 μm and 50 μm (zoom) ( B ) Quantification of IDO1 staining. N = 5 vs 5 vs 4 vs 12 patient samples. C Expression of IDO1 and IFNG in CRC patients with wildtype or mutated DOCK2 . D Expression of IDO1 and IFNG in MSS CRC patients with wildtype or mutated DOCK2 . E Survival plot of TCGA CRC dataset grouped on the mutational status of DOCK2 . Note reduced survival of DOCK2 mutant patients. F Schematic outlining the tryptophan metabolism pathway. Note IDO1 catalysers the conversion of tryptophan to kynurenine. Metabolites altered in Vil Apc Dock2 tumours are highlighted in blue if downregulated (tryptophan) or red if upregulated (xanthurenate). G Volcano plot showing the polar metabolites filtered using p > 0.05 and FC > 1.3. H Individual value plots of log 2 intensity values of selected metabolites. N = 12 vs 12 tumours. Data represented as mean and error bars SD. Statistical analysis for ( B ) was performed using ordinary one-way ANOVA with Tukey’s multiple comparisons. Statistical analysis for ( C ) and ( D ) was performed using student T-test. Statistical analysis for ( H ) was performed using two-tailed Mann–Whitney test. Exact p values are indicated in the panels.

To extend these analyses we utilised the TCGA dataset of sporadic CRC. DOCK2 is mutated in ~10% of cases and in samples where DOCK2 is mutated, expression of both IFNG and IDO1 was significantly higher than non-mutated samples (Fig. 5C ). Importantly, this is also observed in microsatellite stable (MSS) CRC which is generally immune suppressed (Fig. 5D ). We also found that DOCK2 mutation and/or low DOCK2 expression correlates poor prognosis in these patients, further supporting a tumour suppressor role for this protein (Figs. 6E and S6 A, B ).

IDO1 is a tryptophan catabolic enzyme that catalysers the conversion of tryptophan to kynurenine and other downstream metabolites (Fig. 6F ). This pathway plays an important immunomodulatory role by regulating the activity of both regulatory and effector T cells [ 20 , 21 , 22 ]. We next investigated whether the increased IDO1 observed in Dock2 deficient tumours impacted on tryptophan metabolism. To identify metabolic differences between Vil Apc and Vil Apc Dock2 tumours, we performed liquid chromatography mass spectrometry (LC-MS). We extracted metabolites using a solid-liquid biphasic extraction [ 23 ] from 12 tumours of each group and analysed the aqueous fraction by high-resolution HILIC LC-MS [ 24 ]. We identified numerous changes in metabolite levels in Dock2 deficient tumours (Figs. 6G , S6 C, S6D and Table S3 ). Consistent with the elevated expression of IDO1, we observed a depletion of tryptophan following Dock2 deletion (Fig. 6G, H ). Interestingly, we did not observe differences in the levels of kynurenine but the downstream metabolite xanthurenate was elevated in Vil Apc Dock2 tumours suggesting a rapid conversion of kynurenine via this pathway (Fig. 6F–H ).

To further test the functional significance of elevated IDO1 expression and tryptophan metabolism in Dock2 deficient tumours we next examined the effect of IDO1 inhibition on tumourigenesis. We treated Vil Apc Dock2 mice with the IDO1 inhibitor 1-L-MT whilst undergoing repeated rounds of DSS induced colitis (Fig. 7A ). Mice were culled at 47 days post first day of DSS and tissue harvested for analysis. Consistent with a functional role for IDO1 expression in driving IBD-CRC development, mice treated with 1-L-MT had a significantly lower tumour number and tumour burden compared to mice treated with vehicle (Figs. 7B–D and S7A ). Importantly, this was not due to IDO1 inhibition impacting on severity of colitis, confirming the tumour promoting effects of increased epithelial IDO1 activity following Dock2 deletion (Fig. S7B ). Together, our data implicate Dock2 loss of function in the development of IBD-CRC. Mechanistically, this is via immune dysregulation leading to increased infiltration of IFNγ producing T cells, driving IDO1 expression.

figure 7

A Schematic outlining IDO inhibitor (1-L-MT) treatment regimen in Vil Apc Dock2 mice. B Representative H&E-stained colonic swiss rolls, detailing tumour burden in vehicle and IDO1 inhibitor (1-L-MT) treated Vil Apc Dock2 mice (black arrows). Scale bars are 2.5 mm. C Tumour number in vehicle and IDO1 inhibitor treated Vil Apc Dock2 mice. D Total tumour burden in vehicle and IDO1 inhibitor treated Vil Apc Dock2 mice. N = 8 vs 8 mice. E A model outlining the proposed role of Dock2 in suppressing inflammation induced colorectal cancer. Data represented as mean and error bars SD. All statistical analysis for this figure was performed using two-tailed Mann–Whitney test. Exact p values are indicated in the panels.

Taken together, these results suggest that loss of Dock2 leads to immune dysregulation, enhancing tumourigenesis via interferon gamma induced expression of IDO1. In mice lacking Dock2 , IFNγ production is increased in multiple T cell populations, including CD8+ and γδ T cells, and administrating IFNγ to tumour organoids induces robust IDO1 expression suggesting this response is driven by immune cell IFNγ production. Finally, treatment with 1-L-MT, an IDO1 inhibitor, abrogates the increased tumourigenesis observed in mice without Dock2 .

IDO1 is an enzyme catalysing the conversion of tryptophan to kynurenine [ 25 ] and has been directly associated with tumourigenesis, as IDO1+ Paneth cells promote immune evasion in sporadic colorectal cancer [ 26 ]. In colitis-associated cancer, conditional loss of IDO1 in the colonic epithelium with AOM/DSS resulted in fewer colonic tumours [ 18 ]. Additionally, 1-L-MT, an inhibitor of IDO1, decreases proliferation of colon cancer cell lines [ 27 ], as well as reduces tumour burden in AOM/DSS mice [ 27 ]. IDO1 −/− mice treated with AOM/DSS have separately been shown to develop smaller tumours [ 28 ]. We also observed decreased tumourigenesis in a colitis induced Dock2 deficient tumour model following 1-L-MT treatment, outlining the importance of tryptophan metabolism in modulating colitis-induced CRC. IDO1 is associated with a reduction in immune system activity [ 25 ] through several mechanisms. First, macrophages [ 22 ] and dendritic cells [ 21 ] expressing IDO1 suppress T cell proliferation via decreasing the tryptophan pool. Second, dendritic cells expressing IDO1 expand the regulatory T cell (Treg) population thus modulating inflammatory activity [ 20 ]. IDO1 also mediates increased tumourigenesis independently of its effect on T cells, as tumourigenesis was also impaired in IDO1 knock out, immune deficient Rag1 -knockout mice [ 29 ]. Although IDO1 is not abundantly expressed in the colonic epithelium at baseline [ 19 , 30 ], it is expressed in the epithelium during active IBD [ 19 ] consistent with the findings of our study. Interestingly, when overexpressed in the epithelium it results in increased secretory cell differentiation and larger mucus layer in the ileum, as well as reduced sensitivity to DSS [ 31 ]. Along these lines, induction of IDO1 affects colitis severity [ 32 ]. However, the literature varies with respect to IDO1 loss or inhibition, with both positive and negative results reported in colitis: IDO1-deficient mice are less sensitive to DSS colitis [ 33 ], yet mice treated with both IDO1 stimulation and IDO1 inhibitor combined, lose the protective effect of IDO1 stimulation [ 32 ]. Thus, activation of IDO1 activity has the potential to mediate colonic homeostasis and tumourigenesis via multiple mechanisms.

Our data strongly implicate IFNγ signalling in driving expression of IDO1, indeed we observed a robust, dose dependent, transcriptional activation of IDO1 following treatment of colonic organoids with recombinant IFNγ. Subsequent analyses identified increased infiltration of CD3 + , CD8+ and γδ T cells as a source of IFNγ in Dock2 deficient tissue although we cannot rule out other immune populations also playing a role. It is not clear why more IFNγ producing T cells accumulate in tumours formed in Dock2 mutant mice, a gene that is expressed primarily in hematopoietic cells. These cells may be recruited to tumours by increased expression of chemokines or they may proliferate in situ. Neither is it clear why a predominance of IFNγ-producing T cells, normally associated with anti-tumour activity [ 34 , 35 , 36 ], correlate with increased tumourigenesis. For example, Vγ1 γδ T cells play a role in counteracting tumour cell survival through communication with intraepithelial lymphocytes expressing the Vγ7 TCR [ 37 ]. Additionally, IFNγ producing CD8 + T cells are known mediators of anti-tumour immunity [ 38 ]. On the other hand, IFNγ can contribute to immune cell evasion, in particular via modulation of IDO1 expression [ 38 ]. Given that T cells lack Dock2 expression in the model used here, it is tempting to speculate that the absence of DOCK2 suppresses their cancer-killing functions via preventing synapse formation with target cancer cells, a previously described function of DOCK2 [ 39 ]. This defective recognition of cancer cells together with the upregulation of IFNγ in DOCK-2 deficient T cells may result in tumour promotion via increased expression of IDO1. Anti-tumourigenic effects of IDO1, such as induction of T cell apoptosis [ 40 ], were not observed either, arguing against immune escape by T cell ablation. It is reasonable, therefore, to suggest that these increased Dock2 deficient IFNγ producing T cells are dysfunctional. Thus, despite accumulating in the colon they are not sufficient to induce an anti-tumour microenvironment, rather via induction of IDO1 expression, promote one that supports tumour development.

Despite these open questions it appears clear that Dock2 deficiency leads to increased tumourigenesis in an IDO1 dependent manner. Additionally, we found increased levels of IDO1 in IBD-CRC samples and in patients with sporadic CRC carrying mutations in the DOCK2 gene. Furthermore, DOCK2 mutational status was correlated with poor survival outcomes in sporadic CRC. Our identification of this pathway, and its relevance in human disease, could therefore help inform the clinical management of IBD-CRC. With regards to this, it is important first to note the status of IDO1 inhibitors in recent clinical trials. Several trials across multiple solid tumour types have revealed disappointing results [ 41 ]. However, these trials have not selected patients based on IDO1 tumour positivity, and none of these trials have occurred in patients with colorectal cancer, particularly following inflammation. Therefore, it remains to be seen whether IDO1 inhibition has a role in the treatment of inflammation-associated colorectal cancer, and this potential needs to be fully explored. Alternatively, this poor performance may indicate that IDO1 activity is not required to fuel tumour growth in established cancers. In this case, IDO1 inhibitors may be more efficacious when used as chemopreventative agents, preventing the initiation of tumourigenesis. It is worth noting that in our model, we inhibited IDO1 during DSS treatment, suggesting a key role during colonic damage and regeneration cycles. Therefore, IDO1 inhibitors may be particularly beneficial to patients with IBD and at high risk of developing cancer, in particular those carrying previously identified high risk DOCK2 mutations [ 15 ]. Additionally, IDO1 activity, identified through increased xanthurenate, has been shown to negatively correlate with intestinal inflammation itself, and modification upstream of xanthurenate has protected against murine colitis [ 41 ]. Therefore, there are several opportunities for biomarker identification, as well as potential therapeutic agents, in both inflammation and inflammation-associated cancer.

To summarise, we have identified an IDO1 induced tryptophan metabolic pathway regulated by IFNγ producing immune cells driving IBD-CRC, highlighting the intricate role of the immune and interferon gamma response in IBD-CRC (Fig. 7E ). This enhances our understanding of the etiology of IBD-CRC and identifies these processes as potential therapeutic targets for this complex and chronic disease.

Materials and methods

Animal models, mus musculus.

All animal experiments were performed in accordance with a UK Home Office license (Project License 70/8885), and were subject to review by the animal welfare and ethics board of the University of Edinburgh. Mice of both genders were used at an age of 6–12 weeks. Mice were bred at the animal facilities of the University of Edinburgh and were kept in 12 h light–dark cycles and were given access to water and food ad libitum. Mice were maintained in a temperature- (20–26 °C) and humidity- (30–70%) controlled environment. Colonies had a mixed background (50% C57Bl6J, 50% S129). The genetic alleles used for this study were as follows: villin-cre ERT2 [ 42 ], Apc fl [ 43 ] and Dock2 tm1a (EUCOMM). The Dock2 tm1a allele is generated by insertion of a lacZ and neo containing construct into Dock2 gene. Insertion of this construct disrupts gene expression, resulting in a ‘knockout-first’ allele. Therefore, the Dock2 tm1a allele is Dock2 knockout in all mouse tissues, including immune cell populations. Mice were genotyped by Transnetyx (Cordoba, USA). At experiment endpoints, mice were humanely sacrificed by cervical dislocation (CD) in line with UK Home Office regulations.

Animal experiments

Gene deletion in villin-cre ERT2 Apc fl and villin-cre ERT2 Apc fl Dock2 tm1a mice was induced as previously described using intraperitoneal tamoxifen at 80 mg/kg [ 44 ]. Chemical colitis in experimental animals was triggered using DSS (36,000–50,000 Da, MP Biomedicals), reconstituted in distilled water. Mice were treated with two 7-day cycles of 0.5% DSS, with recovery in between, and aged. Labelling of actively replicating cells was achieved through IP injection of 200 μl of BrdU (Amersham Bioscience), 1–2 h prior to Schedule 1 culling. For the inhibitor experiment, mice were treated with 400 mg/kg 1-LMT, an IDO1 inhibitor, daily for the period of DSS administrations, plus two further days via gavage, or vehicle (0.5% methylcellulose/0.5% Tween 80) without drug. Power calculations based on previous results from similar experiments were carried out prior to the experiment to ensure appropriate sample sizes to determine statistically significant effects. Investigators were blinded to mouse genotype during experiment and data analysis.

Patient samples

Anonymised archival paraffin embedded human colonic resection specimens from patients with IBD-associated colorectal cancer held within the Lothian NRS Bioresource were obtained following approval (Sample Requests SR1165 and SR1165- AM01 301120; Ethical Approvals 15/ES/0094 and 20/ES/0061). Informed consent was obtained from all subjects. Relevant areas of normal tissue, inflammation, dysplasia and cancer were identified by a pathologist and samples were used for subsequent immunohistochemical analysis.

Immunohistochemistry

After dewaxing, and where necessary, methacarn-fixed slides were treated with 4% PFA for 10 min. Antigen retrieval at 99 °C was in one of two buffers according to protocol: citrate (pH 6) or EDTA. Hydrogen peroxide was used to prevent endogenous staining at either 1.5% or 3%. Sections then were blocked in 5% goat serum (Sigma) for an hour, before application of the relevant primary antibody at dilutions in 5% goat serum, overnight at 4 °C. The following day the secondary antibody (DAKO Envision) was applied according to manufacturer’s recommendations. Staining for all sections was developed using diaminobenzidine (Thermo Scientific) for three minutes, before counterstaining and mounting. BrdU (BD Transduction (347580)) staining was at 1:500, using a citrate buffer for 25 min heating/30 min cooling, and 10 min hydrogen peroxide 1.5%. CD3 (DAKO (A045229)) was at 1:500, using a EDTA buffer for 15 min heating/30 min cooling, and 10 min hydrogen peroxide 3%. IDO1 (Cell Signalling (51851S)) staining was at 1:100, using a citrate buffer for 20 min heating and 20 min cooling, and 10 min hydrogen peroxide 3%.

Primers used for the amplification of each selected gene are detailed in Table S4 . The reaction used SYBR Select Master Mix (Applied Biosystems), together with forward and reverse primer (combined and diluted 1:10), and water. Gene expression levels were examined in duplicate and normalised to Actb using 2^ -(ΔΔ Cycle threshold) calculations. The protocol involved incubation at 95 °C for 15 min, and then 45 cycles of [denaturation at 95 °C for 10 s, annealing at 60 °C for 30 s, and extension at 72 °C for 30 s]. Primers used for analysis are listed in Table S4 .

RNASeq on individual, isolated tumours was carried out by the Welcome Trust Clinical Research Facility (WTCRF) at the Western General Hospital, Edinburgh. After RNA extraction, samples were checked for RNA quality using a Bioanalyser - all had a RIN of 9.1 or above. Libraries were prepared using 500 ng of each sample with the NEBNEXT Ultra II Directional RNA Library Prep kit (NEB #E7760) and the Poly-A mRNA magnetic isolation module (NEB #E7490). Sequencing was performed using the NextSeq 500/550 High-Output v2.5 Kit (#20024907) on the NextSeq 550 platform (Illumina Inc, #SY-415-1002). All libraries generated greater than 24 million paired end reads. FASTQ files (four lanes per sample) were subsequently uploaded for analysis using DESeq2. Significant genes were considered for pathway analyses using G:profiler ( https://biit.cs.ut.ee/gprofiler/ ). Genes were pre-ranked and analysed for Gene Set Enrichment Analysis (GSEA, v3.0) to generate enrichment plots for Hallmark datasets.

In-situ hybridisation (ISH) staining was performed on 4 µm formalin fixed paraffin embedded sections (FFPE) which had previously been ovened at 60 °C for 2 h. ISH detection for, mRNA probe for Trdc (449358; Advanced Cell Diagnostics) was performed using RNAScope 2.5 LSx (Brown) detection kit (322700; Advanced Cell Diagnostics) on a Leica Bond Rx autostainer strictly according to the manufacturer’s instructions. To complete ISH staining sections were rinsed in tap water, dehydrated through graded ethanols and placed in xylene. The stained sections were coverslipped in xylene using DPX mountant (CellPath, UK).

Western blotting

Protein samples were first denatured in NuPAGE LDS sample loading buffer (ThermoFisher Scientific) together with NuPAGE sample reducing reagent at 99 °C for 5 min, before loading on to a gel. Gels used were SDS (NuPAGE) pre-prepared and were either 4–12% Bis-Tris or 3–8% Tris-Acetate, depending on the size of the protein of interest. Once running was complete, proteins were transferred onto a nylon membrane (PVDF, Amersham) in Nupage transfer buffer containing 20% methanol in ddH 2 0. Membranes were then blocked in 5% milk before applying antibody (IDO1 - Cell Signalling (51851S), or Vinculin - Abcam (ab73412) at 1:1000 overnight at 4 °C. Samples were then washed, before applying secondary antibody at a concentration of 1:5000 for an hour at room temperature, before developing using Pierce ECL (ThermoFisher) kits.

Flow cytometry

1 cm of colon was freshly dissected, 2 cm from the rectum, and digested in a cocktail of 2.5 ml RPMI media (Gibco) containing 5% FCS (in-house) 0.5 mg/ml collagenase (Sigma), 0.5 mg/ml DispaseII (Merck) and 3 mg/ml DNAseI (Roche) in a rocking incubator at 37 °C for 45 min before passing through a 70 μm strainer. 5 × 10 5 cells were plated per sample, with the remainder being used for unstained control. 150 μl of T cell stimulation media (IMDM, Sigma, plus β-mercaptoethanol 50 μM plus Pen-strep (in-house) +T cell stimulation cocktail (1:500, Invitrogen) was added to each well before placing the plate in the 37 °C incubator for three hours. Each stained well was then resuspended in 50 μl blocking buffer (50 μl FACS buffer and 1 μl fc block (TruStain FcX, Biolegend) for 20 min at 4 °C. Meanwhile an antibody mix was prepared using 1 μl per sample of the relevant extracellular antibody diluted 1:50 in Brilliant Stain buffer (BD Biosciences). 50 μl of antibody mix was added to each sample and the plate was further left at 4 °C for 30 min. 100 μl of Zombie live/dead agent (Biolegend) was added at 1:100 in cold PBS and again incubated for 20 min at 4 °C. Samples were resuspended in 75 μl fixation buffer (Invitrogen), again for 20 min at 4 °C. After incubation, 75 μl 1x permeabilisation buffer (Invitrogen, diluted in distilled water) was added before finally resuspending in FACS buffer and keeping at 4 °C overnight.

The following day 100 μl of intracellular antibody mix (diluted 1:100 in permeabilisation buffer) was added to each sample and the plate was incubated at 4 °C for 30 min. After intracellular antibody incubation, 50 μl permeabilisation buffer was added to each well, before resuspending each well in 100 μl FACS buffer. Samples were then flowed on BD Fortessa™. Antibodies used were Zombie Live/Dead (Biolegend (423107) – UV), CD3 (Biolegend (100219) – PE/Cy7), TCRγδ (Biolegend (118127) – AF 488) and IFNγ (Biolegend (505816) – AF647).

Organoid generation and culture

Media for organoid culture used Advanced DMEM/F12 (ADF - Gibco) with 5 mls HEPES (Invitrogen), 5 ml glutamine (in-house) and 5 ml Penstrep (in-house) added. Colonic Wild-type and Dock2tm1a/tm1a organoids were generated directly from epithelial tissue culture as part of this project. Colon was harvested and 25 mM EDTA was added for thirty minutes to catch free magnesium, loosening up the crypts from the surrounding tissue. Once the EDTA reaction was terminated the post-washing supernatant was collected and centrifuged and the end pellet was resuspended in ADF and passed through a 70 μm cell strainer before resuspending in BME and plating. Complete media containing 15% R-spondin-conditioned media (in-house – HEK-293 cells stably expressing HA-R-Spondin), 50 ng/ml EGF, 1X B27 (Life Technologies), 1X N2 (Life Technologies), 1% Noggin-conditioned media (in-house - HEK 293 cells stably expressing Noggin), 50% Wnt3a condition media (in-house –L-cells stably transfected expressing Wnt3a) and 3 μM StemMACS (CHIR99021, Abcam) was used in culture. Established organoids were then passaged a week later. Organoids were treated with mouse interferon-γ (Life Technologies), diluted in media at 1 ng/ml. Protein and RNA were collected form organoids at 24 h after administration of IFNγ, to compare to organoids cultured only in normal media.

Organoid CRISPR

Colonic wild-type and Dock2 tm1a/tm1a organoids were further modified to CRISPR Apc , rendering them Apc −/− and Apc −/− Dock2 tm1a/tm1a . On day 0 cells were split and plated in complete media. On day three they were again split and plated, with the StemMACS concentration raised, now being at 0.6 μl/ml. Additional factors added were Jagged (1 μM, Eurogentec), ROCK inhibitor (10 μM Y-27632, Tocris), and 1 mM VPA (Sigma). On day five viral infection took place, adding 1 ml Accutase™ (Life Technologies) and 4 μl Y-27632 to the harvested pellet. This reaction took place in a water bath at 37 °C for 3 min before termination of the enzymatic reaction with 1 ml 1% BSA and washing in ADF. 5 × 10 5 cells were plated per well of a six well plate with virus at an MOI of 50. Two guide RNAs for Apc , “APC1” (GGATCTGTATCCAGCCGTTC) and “APC3” (AGATCCTTCCCGACTTCCGT), were used. Media for transduction contained EGF, Noggin, R-spondin and Wnt as above, StemMACS 0.6 μl/ml, 1 μM Jagged, 10 μM Y-27632, 1 mM VPA (1:10000), and polybrene (8 μg/ml, Sigma). 24 h later, the infection media was removed and BME overlaid, with fresh media used (containing all constituents of the transduction media bar the polybrene). Media was changed again 24 h after that, this time replacing with ENRW with StemMACS at 1:200 and Y-27632 10 μM only. Media changes were then carried out as necessary until day 11 when growing organoids were split and R-spondin, Wnt and StemMACS removed.

Metabolomics

Polar metabolites were extracted using biphasic extraction. On a 1.5 mL Eppendorf vial, 100 µL of methanol containing internal standard (L-Glutamine_13C5,15N2) was added to the tissue samples, sonicated on ice cold bath for 10 min, following by the addition of 300 uL of MTBE. Vials were shacked for 20 min at 8 °C and 100 µL H2O was added to induce phase separation, followed by vertexing it for one minute and centrifuged at max speed for 10 min at 4 °C. The upper phase containing lipids and the lower phase containing the polar metabolites were individually transferred for new vials. The polar phase was placed at −80 °C for one hour to guarantee protein precipitation, centrifuged at max speed for 10 min at 4 °C and transfer to a 96 well plate for HPLC analysis. Samples were analysed on a Dionex UltiMate 3000 LC System (Thermo Scientific, Waltham, Massachusetts, EUA) coupled to a Q Exactive Orbitrap Mass Spectrometer (Thermo Scientific, Waltham, Massachusetts, EUA) operating in polarity switch mode. Chromatographic separation was achieved using a ZIC®-pHILIC 150 × 2.1 mm column (Merck MilliporeSigma, Burlington, Massachusetts, EUA) using a gradient starting from 20% buffer A (20 mM ammonium carbonate 0.1% ammonium hydroxide solution 25%), and 80% B (acetonitrile) to 80% buffer A, 20% buffer B at 18 min and reconditioning the column to the initial condition until 27.5 min. Mass spectrometry data were processed using Skyline [ 45 ] on a targeted fashion by matching accurate mass and retention time using a in house library acquired from authentic standards. Statistical analysis was performed using metaboanalyst 5.0 [ 46 ].

Diagrams were created using Biorender under a Premium Plan (K.B.M).

Quantification and statistical analysis

Statistical analyses were performed using GraphPad Prism software (v8.3 GraphPad Software, La Jolla, CA, USA) performing the tests as indicated in the figure legends or main text. Significance levels were calculated according: p < 0.05 (*), p < 0.01(**) and p < 0.001 (***). P values are reported on graphs where comparisons are statistically significant. On graphs with multiple comparisons, for clarity, non-significant changes are not shown.

Contact for reagent and resource sharing

Requests for further information, reagents and resources should be directed and will be fulfilled by the Lead Contact, Kevin B Myant: ([email protected]).

Data availability

All relevant data is available from the authors upon request.

Beaugerie L, Itzkowitz SH. Cancers complicating inflammatory bowel disease. N Engl J Med. 2015;372:1441–52.

Article   CAS   PubMed   Google Scholar  

Coskun M, Vermeire S, Nielsen OH. Novel targeted therapies for inflammatory bowel disease. Trends Pharmacol Sci. 2017;38:127–42.

Jewel Samadder N, Valentine JF, Guthery S, Singh H, Bernstein CN, Wan Y, et al. Colorectal cancer in inflammatory bowel diseases: a population-based study in Utah. Dig Dis Sci. 2017;62:2126–32.

Olen O, Erichsen R, Sachs MC, Pedersen L, Halfvarson J, Askling J, et al. Colorectal cancer in ulcerative colitis: a Scandinavian population-based cohort study. Lancet. 2020;395:123–31.

Article   PubMed   Google Scholar  

Cancer Genome Atlas N. Comprehensive molecular characterization of human colon and rectal cancer. Nature. 2012;487:330–7.

Article   Google Scholar  

Din S, Wong K, Mueller MF, Oniscu A, Hewinson J, Black CJ, et al. Mutational analysis identifies therapeutic biomarkers in inflammatory bowel disease-associated colorectal cancers. Clin Cancer Res. 2018;24:5133–42.

Article   CAS   PubMed   PubMed Central   Google Scholar  

Robles AI, Traverso G, Zhang M, Roberts NJ, Khan MA, Joseph C, et al. Whole-exome sequencing analyses of inflammatory bowel disease-associated colorectal cancers. Gastroenterology. 2016;150:931–43.

Gudino V, Cammareri P, Billard CV, Myant KB. Negative regulation of TGFbeta-induced apoptosis by RAC1B enhances intestinal tumourigenesis. Cell Death Dis. 2021;12:873.

Gudino V, Pohl SO, Billard CV, Cammareri P, Bolado A, Aitken S, et al. RAC1B modulates intestinal tumourigenesis via modulation of WNT and EGFR signalling pathways. Nat Commun. 2021;12:2335.

Myant KB, Cammareri P, McGhee EJ, Ridgway RA, Huels DJ, Cordero JB, et al. ROS production and NF-kappaB activation triggered by RAC1 facilitate WNT-driven intestinal stem cell proliferation and colorectal cancer initiation. Cell Stem Cell. 2013;12:761–73.

Parri M, Chiarugi P. Rac and Rho GTPases in cancer cell motility control. Cell Commun Signal. 2010;8:23.

Article   PubMed   PubMed Central   Google Scholar  

Nishihara H, Kobayashi S, Hashimoto Y, Ohba F, Mochizuki N, Kurata T, et al. Non-adherent cell-specific expression of DOCK2, a member of the human CDM-family proteins. Biochim Biophys Acta. 1999;1452:179–87.

Kunimura K, Uruno T, Fukui Y. DOCK family proteins: key players in immune surveillance mechanisms. Int Immunol. 2020;32:5–15.

Ji L, Xu S, Luo H, Zeng F. Insights from DOCK2 in cell function and pathophysiology. Front Mol Biosci. 2022;9:997659.

Peters LA, Perrigoue J, Mortha A, Iuga A, Song WM, Neiman EM, et al. A functional genomics predictive network model identifies regulators of inflammatory bowel disease. Nat Genet. 2017;49:1437–49.

Xie Y, Chen J, Wu B, He T, Xie L, Liu Z. Dock2 affects the host susceptibility to Citrobacter rodentium infection through regulating gut microbiota. Gut Pathog. 2021;13:52.

Myant KB, Cammareri P, Hodder MC, Wills J, Von Kriegsheim A, Gyorffy B, et al. HUWE1 is a critical colonic tumour suppressor gene that prevents MYC signalling, DNA damage accumulation and tumour initiation. EMBO Mol Med. 2017;9:181–97.

Bishnupuri KS, Alvarado DM, Khouri AN, Shabsovich M, Chen B, Dieckgraefe BK, et al. IDO1 and Kynurenine pathway metabolites activate PI3K-Akt signaling in the neoplastic colon epithelium to promote cancer cell proliferation and inhibit apoptosis. Cancer Res. 2019;79:1138–50.

Ciorba MA. Indoleamine 2,3 dioxygenase in intestinal disease. Curr Opin Gastroenterol. 2013;29:146–52.

Chung DJ, Rossi M, Romano E, Ghith J, Yuan J, Munn DH, et al. Indoleamine 2,3-dioxygenase-expressing mature human monocyte-derived dendritic cells expand potent autologous regulatory T cells. Blood. 2009;114:555–63.

Hwu P, Du MX, Lapointe R, Do M, Taylor MW, Young HA. Indoleamine 2,3-dioxygenase production by human dendritic cells results in the inhibition of T cell proliferation. J Immunol. 2000;164:3596–9.

Munn DH, Shafizadeh E, Attwood JT, Bondarev I, Pashine A, Mellor AL. Inhibition of T cell proliferation by macrophage tryptophan catabolism. J Exp Med. 1999;189:1363–72.

Matyash V, Liebisch G, Kurzchalia TV, Shevchenko A, Schwudke D. Lipid extraction by methyl-tert-butyl ether for high-throughput lipidomics. J Lipid Res. 2008;49:1137–46.

Sadiku P, Willson JA, Ryan EM, Sammut D, Coelho P, Watts ER, et al. Neutrophils fuel effective immune responses through gluconeogenesis and glycogenesis. Cell Metab. 2021;33:411–23.

Liu M, Wang X, Wang L, Ma X, Gong Z, Zhang S, et al. Targeting the IDO1 pathway in cancer: from bench to bedside. J Hematol Oncol. 2018;11:100.

Pflugler S, Svinka J, Scharf I, Crncec I, Filipits M, Charoentong P, et al. IDO1(+) Paneth cells promote immune escape of colorectal cancer. Commun Biol. 2020;3:252.

Liu X, Zhou W, Zhang X, Ding Y, Du Q, Hu R. 1-L-MT, an IDO inhibitor, prevented colitis-associated cancer by inducing CDC20 inhibition-mediated mitotic death of colon cancer cells. Int J Cancer. 2018;143:1516–29.

Zhang X, Liu X, Zhou W, Du Q, Yang M, Ding Y, et al. Blockade of IDO-Kynurenine-AhR axis ameliorated colitis-associated colon cancer via inhibiting immune tolerance. Cell Mol Gastroenterol Hepatol. 2021;12:1179–99.

Thaker AI, Rao MS, Bishnupuri KS, Kerr TA, Foster L, Marinshaw JM, et al. IDO1 metabolites activate beta-catenin signaling to promote cancer cell proliferation and colon tumorigenesis in mice. Gastroenterology. 2013;145:416–25.

Theate I, van Baren N, Pilotte L, Moulin P, Larrieu P, Renauld JC, et al. Extensive profiling of the expression of the indoleamine 2,3-dioxygenase 1 protein in normal and tumoral human tissues. Cancer Immunol Res. 2015;3:161–72.

Alvarado DM, Chen B, Iticovici M, Thaker AI, Dai N, VanDussen KL, et al. Epithelial indoleamine 2,3-Dioxygenase 1 Modulates Aryl hydrocarbon receptor and notch signaling to increase differentiation of secretory cells and alter mucus-associated microbiota. Gastroenterology. 2019;157:1093–108.

Ciorba MA, Bettonville EE, McDonald KG, Metz R, Prendergast GC, Newberry RD, et al. Induction of IDO-1 by immunostimulatory DNA limits severity of experimental colitis. J Immunol. 2010;184:3907–16.

Shon WJ, Lee YK, Shin JH, Choi EY, Shin DM. Severity of DSS-induced colitis is reduced in Ido1-deficient mice with down-regulation of TLR-MyD88-NF-kB transcriptional networks. Sci Rep. 2015;5:17305.

Gao Y, Yang W, Pan M, Scully E, Girardi M, Augenlicht LH, et al. Gamma delta T cells provide an early source of interferon gamma in tumor immunity. J Exp Med. 2003;198:433–42.

He D, Li H, Yusuf N, Elmets CA, Li J, Mountz JD, et al. IL-17 promotes tumor development through the induction of tumor promoting microenvironments at tumor sites and myeloid-derived suppressor cells. J Immunol. 2010;184:2281–8.

Liu Z, Eltoum IE, Guo B, Beck BH, Cloud GA, Lopez RD. Protective immunosurveillance and therapeutic antitumor activity of gammadelta T cells demonstrated in a mouse model of prostate cancer. J Immunol. 2008;180:6044–53.

Reis BS, Darcy PW, Khan IZ, Moon CS, Kornberg AE, Schneider VS, et al. TCR-Vgammadelta usage distinguishes protumor from antitumor intestinal gammadelta T cell subsets. Science. 2022;377:276–84.

Du W, Frankel TL, Green M, Zou W. IFNgamma signaling integrity in colorectal cancer immunity and immunotherapy. Cell Mol Immunol. 2022;19:23–32.

Sanui T, Inayoshi A, Noda M, Iwata E, Stein JV, Sasazuki T, et al. DOCK2 regulates Rac activation and cytoskeletal reorganization through interaction with ELMO1. Blood. 2003;102:2948–50.

Gocher AM, Workman CJ, Vignali DAA. Interferon-gamma: teammate or opponent in the tumour microenvironment? Nat Rev Immunol. 2022;22:158–72.

Fujiwara Y, Kato S, Nesline MK, Conroy JM, DePietro P, Pabla S, et al. Indoleamine 2,3-dioxygenase (IDO) inhibitors and cancer immunotherapy. Cancer Treat Rev. 2022;110:102461.

el Marjou F, Janssen KP, Chang BH, Li M, Hindie V, Chan L, et al. Tissue-specific and inducible Cre-mediated recombination in the gut epithelium. Genesis. 2004;39:186–93.

Shibata H, Toyama K, Shioya H, Ito M, Hirota M, Hasegawa S, et al. Rapid colorectal adenoma formation initiated by conditional targeting of the Apc gene. Science. 1997;278:120–3.

Hall AE, Pohl SO, Cammareri P, Aitken S, Younger NT, Raponi M, et al. RNA splicing is a key mediator of tumour cell plasticity and a therapeutic vulnerability in colorectal cancer. Nat Commun. 2022;13:2791.

MacLean B, Tomazela DM, Shulman N, Chambers M, Finney GL, Frewen B, et al. Skyline: an open source document editor for creating and analyzing targeted proteomics experiments. Bioinformatics. 2010;26:966–8.

Xia J, Psychogios N, Young N, Wishart DS. MetaboAnalyst: a web server for metabolomic data analysis and interpretation. Nucleic Acids Res. 2009;37:W652–60.

Download references

Acknowledgements

This work was funded by a Wellcome Trust clinical research fellowship through the Edinburgh Clinical Academic Track (ECAT) programme, 214369/Z/18/Z (AMDC), as well as Cancer Research UK (CRUK) under Career Development Fellowship, A19166 (KBM) and Small Molecule Drug Discovery Project Award, A25808 (KBM), the European Research Council under Starting Grant, COLGENES – 715782 (KBM), Marie Skłodowska Curie Actions European Fellowship (GDCOLCA 800112; to TS); Naito Foundation Grant for Research Abroad (to TS) and Career Establishment Award RCCCEA-Nov21\100003; to SB. Coffelt. Metabolomics was supported by the Wellcome Trust (Multiuser Equipment Grant 208402/Z/17/Z). We thank the University of Edinburgh’s Institute of Genetics and Cancer (IGC) technical staff for providing support for some of the experiments and we thank the animal technicians at the Biomedical Research Facility (BRF) facility for animal husbandry support. We thank Prof Owen Sansom for providing us with the villin-cre ERT2 and Apc fl mouse lines. Finally, we thank Lothian NRS BioResource for access to human pathological slides. SD acknowledges the support of NHS Research Scotland via NHS Lothian.

Author information

Authors and affiliations.

Institute of Genetics and Cancer, The University of Edinburgh, Western General Hospital Campus, Edinburgh, UK

Antonia M. D. Churchhouse, Caroline V. Billard, Sebastian Ö. G. Pohl, Nora J. Doleschall, Kevin Donnelly, Mark J. Arends, Jair Marques Junior, Alex Von Kriegsheim & Kevin B. Myant

Cancer Research UK Scotland Institute, Garscube Estate, Glasgow, UK

Toshiyasu Suzuki, Colin Nixon & Seth B. Coffelt

School of Cancer Sciences, University of Glasgow, Glasgow, UK

Toshiyasu Suzuki & Seth B. Coffelt

Edinburgh IBD Unit, Western General Hospital, Edinburgh, UK

Shahida Din

Department of Pathology, Western General Hospital, Edinburgh, UK

Kathryn Kirkwood

You can also search for this author in PubMed   Google Scholar

Contributions

AMDC helped with study design and carried out the majority of experiments. CVB provided technical support for animal work, carried out some animal experiments and carried out histological analysis. SOGP carried out bioinformatics analysis of RNAseq data and investigated IDO1 induction by IFNγ. KD carried out bioinformatics analysis of RNAseq data. TS helped with study design for immune analysis and carry out flow cytometry analysis of immune cell populations and discussed the results and contributed to the final manuscript. NJD carried out the coIF experiments. MJA, SD, and KK provided human tissue samples and assistance with histological analysis alongside Colin Nixon. JMJ and AVK carried out the metabolomic analysis. SBC provided assistance with immune cell characterisation and analysis and interpretation of results. KBM designed the project and helped support the experiments. AMDC, SBC, and KBM wrote the manuscript.

Corresponding author

Correspondence to Kevin B. Myant .

Ethics declarations

Competing interests.

The authors declare no competing interests.

Ethics approval and consent to participate

All methods were performed in accordance with relevant guidelines and regulations. Approval for animal experiments has been obtained from the animal welfare and ethics board of the University of Edinburgh and all animal experiments were performed in accordance with a UK Home Office license (Project License 70/8885). Informed consent was obtained from all human subjects providing samples for analysis (Sample Requests SR1165 and SR1165- AM01 301120; Ethical Approvals 15/ES/0094 and 20/ES/0061).

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Supplemental figures and legends, rights and permissions.

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Cite this article.

Churchhouse, A.M.D., Billard, C.V., Suzuki, T. et al. Loss of DOCK2 potentiates Inflammatory Bowel Disease–associated colorectal cancer via immune dysfunction and IFNγ induction of IDO1 expression. Oncogene (2024). https://doi.org/10.1038/s41388-024-03135-9

Download citation

Received : 08 August 2023

Revised : 05 August 2024

Accepted : 13 August 2024

Published : 07 September 2024

DOI : https://doi.org/10.1038/s41388-024-03135-9

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

Quick links

  • Explore articles by subject
  • Guide to authors
  • Editorial policies

non directional two tailed hypothesis

IMAGES

  1. Two-Tailed Test in Hypothesis Testing || Non-Directional Test ||

    non directional two tailed hypothesis

  2. What Is a Two-Tailed Test? Definition and Example

    non directional two tailed hypothesis

  3. PPT

    non directional two tailed hypothesis

  4. Hypothesis Testing

    non directional two tailed hypothesis

  5. Hypotheses.

    non directional two tailed hypothesis

  6. Hypothesis Testing: Upper, Lower, and Two Tailed Tests

    non directional two tailed hypothesis

VIDEO

  1. 1 tailed and 2 tailed Hypothesis

  2. One tailed hypothesis and two tailed hypothesis

  3. Chapter 09: One sample hypothesis testing-worked examples

  4. One tailed hypothesis #psychology#statistics #ugcnet #netpsychology

  5. Conducting a Two Tailed Hypothesis for Difference Means Using Excel or Google Sheets

  6. Hypothesis Meaning in Bengali

COMMENTS

  1. Directional vs Non-Directional Hypothesis: Key Difference

    A non-directional hypothesis, also known as a two-tailed hypothesis, is appropriate to use in several situations where researchers do not have a specific prediction about the direction of the relationship or difference between variables. Here are some scenarios where a non-directional hypothesis is commonly employed: Exploratory Research:

  2. Hypotheses; directional and non-directional

    The directional hypothesis can also state a negative correlation, e.g. the higher the number of face-book friends, the lower the life satisfaction score ". Non-directional hypothesis: A non-directional (or two tailed hypothesis) simply states that there will be a difference between the two groups/conditions but does not say which will be ...

  3. Directional and non-directional hypothesis: A Comprehensive Guide

    Non-directional hypotheses, also known as two-tailed hypotheses, are statements in research that indicate the presence of a relationship or difference between variables without specifying the direction of the effect. Instead of making predictions about the specific direction of the relationship or difference, non-directional hypotheses simply ...

  4. One-Tailed and Two-Tailed Hypothesis Tests Explained

    One-Tailed and Two-Tailed Hypothesis Tests Explained

  5. Data analysis: hypothesis testing: 4.2 Two-tailed tests

    Non-directional hypotheses are also known as two-tailed hypotheses. The term 'two-tailed' comes from the fact that the statistical test used to evaluate the hypothesis is based on the assumption that the difference or relationship could occur in either direction, resulting in two 'tails' in the probability distribution.

  6. Research Hypothesis In Psychology: Types, & Examples

    Research Hypothesis In Psychology: Types, & Examples

  7. 5.2

    5.2 - Writing Hypotheses. The first step in conducting a hypothesis test is to write the hypothesis statements that are going to be tested. For each test you will have a null hypothesis (H 0) and an alternative hypothesis (H a). When writing hypotheses there are three things that we need to know: (1) the parameter that we are testing (2) the ...

  8. Hypothesis Testing: Upper-, Lower, and Two Tailed Tests

    Hypothesis Testing: Upper-, Lower, and Two Tailed Tests

  9. Understanding Statistical Testing

    Directional hypotheses are sometimes introduced as a point hypothesis and a directional hypothesis such as H 0: θ = 0 versus H 1: θ > 0 (Neutens & Rubinson, ... For example, the common practice of treating insignificant results of a formal two-tailed test as evidence that there is no effect should be avoided. Instead, it should be ...

  10. PDF Chapter 6: Research methods Hypotheses: directional or non-directional

    If a hypothesis does not state a direction but simply says that one factor affects another, or that there is an association or correlation between two variables then it is called a non-directional (two-tailed) hypothesis. Distinguishing between directional and non-directional hypotheses is really very straightforward but be careful! Sometimes ...

  11. Two-Tailed Hypothesis Tests: 3 Example Problems

    HA (Alternative Hypothesis): μ ≠ 10 inches. This is an example of a two-tailed hypothesis test because the alternative hypothesis contains the not equal "≠" sign. The botanist believes that the new fertilizer will influence plant growth, but doesn't specify whether it will cause average growth to increase or decrease.

  12. Directional vs. Non-Directional Hypothesis in Research

    A non-directional hypothesis, also known as a two-tailed hypothesis, does not make a specific prediction about the direction of the relationship between variables. Instead, it states that there is a relationship, but without indicating whether it will be positive or negative.

  13. One-Tailed vs. Two-Tailed Tests

    One-Tailed vs. Two-Tailed Tests | Overview & Examples

  14. Hypothesis

    A non-directional experimental hypothesis (also known as two-tailed) does not predict the direction of the change/difference (it is an 'open goal' i.e. anything could happen) A non-directional hypothesis is usually used when there is either no or little previous research which support a particular theory or outcome i.e. what the researcher ...

  15. 8.3.2

    Type of Hypothesis Test: Two-tailed, non-directional: Right-tailed, directional: ... This is a two-tailed test, so we need to take into account both the left and right sides of the curve. \(p=0.4510+0.4510=0.9020\) Step 4: Make a decision. We will compare our \(p\)-value from step 3 to a standard alpha level of 0.05.

  16. Non-Directional Hypothesis

    A non-directional hypothesis is a two-tailed hypothesis that does not predict the direction of the difference or relationship (e.g. girls and boys are different in terms of helpfulness). Share on Facebook; Share on Twitter; Share by Email; Research Methods: MCQ Revision Test 1 for AQA A Level Psychology

  17. What is a Directional Hypothesis? (Definition & Examples)

    Note that directional hypothesis tests are also called "one-tailed" tests and non-directional hypothesis tests are also called "two-tailed" tests. Check out the following examples to gain a better understanding of directional vs. non-directional hypothesis tests. Example 1: Baseball Programs. A baseball coach believes a certain 4-week ...

  18. One-tailed and Two-tailed Tests

    The long answer is: Use one tailed tests when you have a specific hypothesis about the direction of your relationship. Some examples include you hypothesize that one group mean is larger than the other; you hypothesize that the correlation is positive; you hypothesize that the proportion is below .5. The short answer is: Never use one tailed tests.

  19. Fisher, Neyman-Pearson or NHST? A tutorial for teaching data testing

    One-tailed and two-tailed tests. With some tests (e.g., F-tests) research data can only be tested against one side of the null distribution (one-tailed tests), ... The tail of a test depends on the test in question, not on whether the null hypothesis is directional or non-directional. However, you can use two-tailed tests as one-tailed ones ...

  20. 8.1.2

    Type of Hypothesis Test: Two-tailed, non-directional: Right-tailed, directional: Left-tailed, directional: Where \(p_0\) is the hypothesized population proportion that you are comparing your sample to. 2. Calculate the test statistic. When using the normal approximation method we will be using a z test statistic. The z test statistic tells us ...

  21. Is two-tailed testing for directional research hypotheses tests

    1. Introduction. Standard textbooks on statistics clearly state that non-directional research hypotheses should be tested using two-tailed testing while one-tailed testing is appropriate for testing directional research hypotheses (e.g., Churchill and Iacobucci, 2002, Pfaffenberger and Patterson, 1987).However, during the actual conduct of statistical testing, this advice is not often heeded.

  22. 8.1: The null and alternative hypotheses

    The Null hypothesis \(\left(H_{O}\right)\) is a statement about the comparisons, e.g., between a sample statistic and the population, or between two treatment groups. The former is referred to as a one-tailed test whereas the latter is called a two-tailed test. The null hypothesis is typically "no statistical difference" between the ...

  23. Z-8: Two-Sample and Directional Hypothesis Testing

    Directional vs. non-directional testing. A two-tailed test for our mice; A one-tailed test for our mice; Critical t-values for one-tailed and two-tailed tests; ... Directional hypothesis testing vs non-directional testing. Remember the example of testing the effect of antibiotics on mice in Lesson 7. The point of the study was to find out if ...

  24. Hypotheses AO1 AO2

    A non-directional (2-tailed) hypothesis only has to predict there will be a difference in the scores between two groups - not which group will do best. For example, Schmolck et al. (2002) weren't sure whether H.M. would do better or worse at semantic menmory tests than the other MTL patients.

  25. Brain organoid maturation and implantation integration based on

    (U) A network plot of the functional connections within the cortical organoids is shown, including the connectivity between the two rows and the different electrode points in the same row of electrodes. Mean ± SD is shown for each condition. *P<0.05, **P<0.01, and ***P<0.001 in unpaired two-tailed t-test. Electrical stimulation altered the ...

  26. Loss of DOCK2 potentiates Inflammatory Bowel Disease ...

    Together, this supports the hypothesis that increased IFNγ production, from non-cell autonomous sources, is driving elevated IDO1 expression in tumour epithelial cells in Dock2 deficient mice ...