• Bipolar Disorder
  • Therapy Center
  • When To See a Therapist
  • Types of Therapy
  • Best Online Therapy
  • Best Couples Therapy
  • Managing Stress
  • Sleep and Dreaming
  • Understanding Emotions
  • Self-Improvement
  • Healthy Relationships
  • Student Resources
  • Personality Types
  • Sweepstakes
  • Guided Meditations
  • Verywell Mind Insights
  • 2024 Verywell Mind 25
  • Mental Health in the Classroom
  • Editorial Process
  • Meet Our Review Board
  • Crisis Support

The Random Selection Experiment Method

When researchers need to select a representative sample from a larger population, they often utilize a method known as random selection. In this selection process, each member of a group stands an equal chance of being chosen as a participant in the study.

Random Selection vs. Random Assignment

How does random selection differ from  random assignment ? Random selection refers to how the sample is drawn from the population as a whole, whereas random assignment refers to how the participants are then assigned to either the experimental or control groups.

It is possible to have both random selection and random assignment in an experiment.

Imagine that you use random selection to draw 500 people from a population to participate in your study. You then use random assignment to assign 250 of your participants to a control group (the group that does not receive the treatment or independent variable) and you assign 250 of the participants to the experimental group (the group that receives the treatment or independent variable).

Why do researchers utilize random selection? The purpose is to increase the generalizability of the results.

By drawing a random sample from a larger population, the goal is that the sample will be representative of the larger group and less likely to be subject to bias.

Factors Involved

Imagine a researcher is selecting people to participate in a study. To pick participants, they may choose people using a technique that is the statistical equivalent of a coin toss.

They may begin by using random selection to pick geographic regions from which to draw participants. They may then use the same selection process to pick cities, neighborhoods, households, age ranges, and individual participants.

Another important thing to remember is that larger sample sizes tend to be more representative. Even random selection can lead to a biased or limited sample if the sample size is small.

When the sample size is small, an unusual participant can have an undue influence over the sample as a whole. Using a larger sample size tends to dilute the effects of unusual participants and prevent them from skewing the results.

Lin L.  Bias caused by sampling error in meta-analysis with small sample sizes .  PLoS ONE . 2018;13(9):e0204056. doi:10.1371/journal.pone.0204056

Elmes DG, Kantowitz BH, Roediger HL.  Research Methods in Psychology. Belmont, CA: Wadsworth; 2012.

By Kendra Cherry, MSEd Kendra Cherry, MS, is a psychosocial rehabilitation specialist, psychology educator, and author of the "Everything Psychology Book."

Explore Psychology

What Is Random Selection?

Categories Dictionary

Random selection refers to a process that researchers use to pick participants for a study. When using this method, every single member of a population has an equal chance of being chosen as a subject.

This process is an important research tool used in psychology research, allowing scientists to create representative samples from which conclusions can be drawn and applied to the larger population.

Table of Contents

Examples of Random Selection

Random selection is a crucial technique in psychology research to ensure that samples are representative of the population, thus enhancing the generalizability of the findings. Here are a few brief examples of how random selection can be used in different areas of psychology research:

Survey Research on Mental Health

A researcher wants to study the prevalence of anxiety disorders among adults in a city. They use random selection to choose a sample of adults from the city’s population registry. This ensures that every adult in the city has an equal chance of being selected, making the sample representative of the entire adult population.

Experimental Research on Cognitive Processes:

To investigate the effects of sleep deprivation on memory, a researcher randomly selects participants from a university’s student population. By randomly assigning these students to either a sleep-deprived group or a control group, the researcher ensures that any differences in memory performance are likely due to the manipulation of sleep rather than pre-existing differences between groups.

Developmental Psychology Studies

A study aims to understand the development of language skills in toddlers. The researcher randomly selects toddlers from several daycare centers in a region. This random selection helps ensure that the sample includes children from diverse backgrounds, leading to more generalizable findings about language development.

Clinical Trials for Psychological Interventions

In testing a new therapeutic intervention for depression, a researcher randomly selects participants from a pool of individuals diagnosed with depression. Participants are then randomly assigned to either the intervention group or a control group (e.g., receiving standard care). This random selection and assignment help control for potential confounding variables and biases.

Social Psychology Research

To study the impact of group dynamics on decision-making, a researcher randomly selects employees from different departments of a large corporation. By using random selection, the researcher can ensure that the sample is not biased towards any particular department, making the findings more applicable across the entire corporation.

These examples illustrate how random selection helps create representative samples and enhances the internal and external validity of psychological research.

Random Selection vs. Random Assignment

It is important to note that random selection is not the same as random assignment . While random selection involves how participants are chosen for a study, random assignment involves how those chosen are then assigned to different groups in the experiment.

Many studies and experiments actually use both random selection and random assignment.

For example, random selection might be used to draw 100 students to participate in a study. Each of these 100 participants would then be randomly assigned to either the control group or the experimental group.

Reasons to Use Random Selection

What is the reason that researchers choose to use random selection when conducting research?

Some key reasons include:

Increased Generalizability

Random selection is one way to help improve the generalizability of the results. A sample is drawn from a larger population. Researchers want to be sure that the sample they use in their study accurately reflects the characteristics of the larger group.

The more representative the sample is, the better able the researchers can generalize the results of their experiment to a larger population.

By randomly selecting participants for a study, researchers can also help minimize the possibility of bias influencing the results.

Reduced Outlier Effects

Random selection helps ensure that anomalies will not skew results. By randomly selecting participants for a study, researchers are less likely to draw on subjects that may share unusual characteristics in common.

For example, suppose researchers were interested in learning how many people in the general population are left-handed. In that case, the results might be skewed if subjects were inadvertently drawn from a group that included an unusually high number of left-handed individuals.

Random selection ensures that the group better represents what exists in the real world.

Hilbert, S. (2017). Random selection . In: Zeigler-Hill, V., Shackelford, T. (eds) Encyclopedia of Personality and Individual Differences . Springer, Cham. https://doi.org/10.1007/978-3-319-28099-8_1344-1

Martínez-Mesa, J., González-Chica, D. A., Duquia, R. P., Bonamigo, R. R., & Bastos, J. L. (2016). Sampling: how to select participants in my research study ?  Anais brasileiros de dermatologia ,  91 (3), 326–330. https://doi.org/10.1590/abd1806-4841.20165254

randomly selected in research

Sampling Methods & Strategies 101

Everything you need to know (including examples)

By: Derek Jansen (MBA) | Expert Reviewed By: Kerryn Warren (PhD) | January 2023

If you’re new to research, sooner or later you’re bound to wander into the intimidating world of sampling methods and strategies. If you find yourself on this page, chances are you’re feeling a little overwhelmed or confused. Fear not – in this post we’ll unpack sampling in straightforward language , along with loads of examples .

Overview: Sampling Methods & Strategies

  • What is sampling in a research context?
  • The two overarching approaches

Simple random sampling

Stratified random sampling, cluster sampling, systematic sampling, purposive sampling, convenience sampling, snowball sampling.

  • How to choose the right sampling method

What (exactly) is sampling?

At the simplest level, sampling (within a research context) is the process of selecting a subset of participants from a larger group . For example, if your research involved assessing US consumers’ perceptions about a particular brand of laundry detergent, you wouldn’t be able to collect data from every single person that uses laundry detergent (good luck with that!) – but you could potentially collect data from a smaller subset of this group.

In technical terms, the larger group is referred to as the population , and the subset (the group you’ll actually engage with in your research) is called the sample . Put another way, you can look at the population as a full cake and the sample as a single slice of that cake. In an ideal world, you’d want your sample to be perfectly representative of the population, as that would allow you to generalise your findings to the entire population. In other words, you’d want to cut a perfect cross-sectional slice of cake, such that the slice reflects every layer of the cake in perfect proportion.

Achieving a truly representative sample is, unfortunately, a little trickier than slicing a cake, as there are many practical challenges and obstacles to achieving this in a real-world setting. Thankfully though, you don’t always need to have a perfectly representative sample – it all depends on the specific research aims of each study – so don’t stress yourself out about that just yet!

With the concept of sampling broadly defined, let’s look at the different approaches to sampling to get a better understanding of what it all looks like in practice.

randomly selected in research

The two overarching sampling approaches

At the highest level, there are two approaches to sampling: probability sampling and non-probability sampling . Within each of these, there are a variety of sampling methods , which we’ll explore a little later.

Probability sampling involves selecting participants (or any unit of interest) on a statistically random basis , which is why it’s also called “random sampling”. In other words, the selection of each individual participant is based on a pre-determined process (not the discretion of the researcher). As a result, this approach achieves a random sample.

Probability-based sampling methods are most commonly used in quantitative research , especially when it’s important to achieve a representative sample that allows the researcher to generalise their findings.

Non-probability sampling , on the other hand, refers to sampling methods in which the selection of participants is not statistically random . In other words, the selection of individual participants is based on the discretion and judgment of the researcher, rather than on a pre-determined process.

Non-probability sampling methods are commonly used in qualitative research , where the richness and depth of the data are more important than the generalisability of the findings.

If that all sounds a little too conceptual and fluffy, don’t worry. Let’s take a look at some actual sampling methods to make it more tangible.

Need a helping hand?

randomly selected in research

Probability-based sampling methods

First, we’ll look at four common probability-based (random) sampling methods:

Importantly, this is not a comprehensive list of all the probability sampling methods – these are just four of the most common ones. So, if you’re interested in adopting a probability-based sampling approach, be sure to explore all the options.

Simple random sampling involves selecting participants in a completely random fashion , where each participant has an equal chance of being selected. Basically, this sampling method is the equivalent of pulling names out of a hat , except that you can do it digitally. For example, if you had a list of 500 people, you could use a random number generator to draw a list of 50 numbers (each number, reflecting a participant) and then use that dataset as your sample.

Thanks to its simplicity, simple random sampling is easy to implement , and as a consequence, is typically quite cheap and efficient . Given that the selection process is completely random, the results can be generalised fairly reliably. However, this also means it can hide the impact of large subgroups within the data, which can result in minority subgroups having little representation in the results – if any at all. To address this, one needs to take a slightly different approach, which we’ll look at next.

Stratified random sampling is similar to simple random sampling, but it kicks things up a notch. As the name suggests, stratified sampling involves selecting participants randomly , but from within certain pre-defined subgroups (i.e., strata) that share a common trait . For example, you might divide the population into strata based on gender, ethnicity, age range or level of education, and then select randomly from each group.

The benefit of this sampling method is that it gives you more control over the impact of large subgroups (strata) within the population. For example, if a population comprises 80% males and 20% females, you may want to “balance” this skew out by selecting a random sample from an equal number of males and females. This would, of course, reduce the representativeness of the sample, but it would allow you to identify differences between subgroups. So, depending on your research aims, the stratified approach could work well.

Free Webinar: Research Methodology 101

Next on the list is cluster sampling. As the name suggests, this sampling method involves sampling from naturally occurring, mutually exclusive clusters within a population – for example, area codes within a city or cities within a country. Once the clusters are defined, a set of clusters are randomly selected and then a set of participants are randomly selected from each cluster.

Now, you’re probably wondering, “how is cluster sampling different from stratified random sampling?”. Well, let’s look at the previous example where each cluster reflects an area code in a given city.

With cluster sampling, you would collect data from clusters of participants in a handful of area codes (let’s say 5 neighbourhoods). Conversely, with stratified random sampling, you would need to collect data from all over the city (i.e., many more neighbourhoods). You’d still achieve the same sample size either way (let’s say 200 people, for example), but with stratified sampling, you’d need to do a lot more running around, as participants would be scattered across a vast geographic area. As a result, cluster sampling is often the more practical and economical option.

If that all sounds a little mind-bending, you can use the following general rule of thumb. If a population is relatively homogeneous , cluster sampling will often be adequate. Conversely, if a population is quite heterogeneous (i.e., diverse), stratified sampling will generally be more appropriate.

The last probability sampling method we’ll look at is systematic sampling. This method simply involves selecting participants at a set interval , starting from a random point .

For example, if you have a list of students that reflects the population of a university, you could systematically sample that population by selecting participants at an interval of 8 . In other words, you would randomly select a starting point – let’s say student number 40 – followed by student 48, 56, 64, etc.

What’s important with systematic sampling is that the population list you select from needs to be randomly ordered . If there are underlying patterns in the list (for example, if the list is ordered by gender, IQ, age, etc.), this will result in a non-random sample, which would defeat the purpose of adopting this sampling method. Of course, you could safeguard against this by “shuffling” your population list using a random number generator or similar tool.

Systematic sampling simply involves selecting participants at a set interval (e.g., every 10th person), starting from a random point.

Non-probability-based sampling methods

Right, now that we’ve looked at a few probability-based sampling methods, let’s look at three non-probability methods :

Again, this is not an exhaustive list of all possible sampling methods, so be sure to explore further if you’re interested in adopting a non-probability sampling approach.

First up, we’ve got purposive sampling – also known as judgment , selective or subjective sampling. Again, the name provides some clues, as this method involves the researcher selecting participants using his or her own judgement , based on the purpose of the study (i.e., the research aims).

For example, suppose your research aims were to understand the perceptions of hyper-loyal customers of a particular retail store. In that case, you could use your judgement to engage with frequent shoppers, as well as rare or occasional shoppers, to understand what judgements drive the two behavioural extremes .

Purposive sampling is often used in studies where the aim is to gather information from a small population (especially rare or hard-to-find populations), as it allows the researcher to target specific individuals who have unique knowledge or experience . Naturally, this sampling method is quite prone to researcher bias and judgement error, and it’s unlikely to produce generalisable results, so it’s best suited to studies where the aim is to go deep rather than broad .

Purposive sampling involves the researcher selecting participants using their own judgement, based on the purpose of the study.

Next up, we have convenience sampling. As the name suggests, with this method, participants are selected based on their availability or accessibility . In other words, the sample is selected based on how convenient it is for the researcher to access it, as opposed to using a defined and objective process.

Naturally, convenience sampling provides a quick and easy way to gather data, as the sample is selected based on the individuals who are readily available or willing to participate. This makes it an attractive option if you’re particularly tight on resources and/or time. However, as you’d expect, this sampling method is unlikely to produce a representative sample and will of course be vulnerable to researcher bias , so it’s important to approach it with caution.

Last but not least, we have the snowball sampling method. This method relies on referrals from initial participants to recruit additional participants. In other words, the initial subjects form the first (small) snowball and each additional subject recruited through referral is added to the snowball, making it larger as it rolls along .

Snowball sampling is often used in research contexts where it’s difficult to identify and access a particular population. For example, people with a rare medical condition or members of an exclusive group. It can also be useful in cases where the research topic is sensitive or taboo and people are unlikely to open up unless they’re referred by someone they trust.

Simply put, snowball sampling is ideal for research that involves reaching hard-to-access populations . But, keep in mind that, once again, it’s a sampling method that’s highly prone to researcher bias and is unlikely to produce a representative sample. So, make sure that it aligns with your research aims and questions before adopting this method.

How to choose a sampling method

Now that we’ve looked at a few popular sampling methods (both probability and non-probability based), the obvious question is, “ how do I choose the right sampling method for my study?”. When selecting a sampling method for your research project, you’ll need to consider two important factors: your research aims and your resources .

As with all research design and methodology choices, your sampling approach needs to be guided by and aligned with your research aims, objectives and research questions – in other words, your golden thread. Specifically, you need to consider whether your research aims are primarily concerned with producing generalisable findings (in which case, you’ll likely opt for a probability-based sampling method) or with achieving rich , deep insights (in which case, a non-probability-based approach could be more practical). Typically, quantitative studies lean toward the former, while qualitative studies aim for the latter, so be sure to consider your broader methodology as well.

The second factor you need to consider is your resources and, more generally, the practical constraints at play. If, for example, you have easy, free access to a large sample at your workplace or university and a healthy budget to help you attract participants, that will open up multiple options in terms of sampling methods. Conversely, if you’re cash-strapped, short on time and don’t have unfettered access to your population of interest, you may be restricted to convenience or referral-based methods.

In short, be ready for trade-offs – you won’t always be able to utilise the “perfect” sampling method for your study, and that’s okay. Much like all the other methodological choices you’ll make as part of your study, you’ll often need to compromise and accept practical trade-offs when it comes to sampling. Don’t let this get you down though – as long as your sampling choice is well explained and justified, and the limitations of your approach are clearly articulated, you’ll be on the right track.

randomly selected in research

Let’s recap…

In this post, we’ve covered the basics of sampling within the context of a typical research project.

  • Sampling refers to the process of defining a subgroup (sample) from the larger group of interest (population).
  • The two overarching approaches to sampling are probability sampling (random) and non-probability sampling .
  • Common probability-based sampling methods include simple random sampling, stratified random sampling, cluster sampling and systematic sampling.
  • Common non-probability-based sampling methods include purposive sampling, convenience sampling and snowball sampling.
  • When choosing a sampling method, you need to consider your research aims , objectives and questions, as well as your resources and other practical constraints .

If you’d like to see an example of a sampling strategy in action, be sure to check out our research methodology chapter sample .

Last but not least, if you need hands-on help with your sampling (or any other aspect of your research), take a look at our 1-on-1 coaching service , where we guide you through each step of the research process, at your own pace.

randomly selected in research

Psst... there’s more!

This post was based on one of our popular Research Bootcamps . If you're working on a research project, you'll definitely want to check this out ...

Abby

Excellent and helpful. Best site to get a full understanding of Research methodology. I’m nolonger as “clueless “..😉

Takele Gezaheg Demie

Excellent and helpful for junior researcher!

Andrea

Grad Coach tutorials are excellent – I recommend them to everyone doing research. I will be working with a sample of imprisoned women and now have a much clearer idea concerning sampling. Thank you to all at Grad Coach for generously sharing your expertise with students.

Submit a Comment Cancel reply

Your email address will not be published. Required fields are marked *

Save my name, email, and website in this browser for the next time I comment.

  • Print Friendly

Difference between Random Selection and Random Assignment

Random selection and random assignment are commonly confused or used interchangeably, though the terms refer to entirely different processes.  Random selection refers to how sample members (study participants) are selected from the population for inclusion in the study.  Random assignment is an aspect of experimental design in which study participants are assigned to the treatment or control group using a random procedure.

Random selection requires the use of some form of random sampling (such as stratified random sampling , in which the population is sorted into groups from which sample members are chosen randomly).  Random sampling is a probability sampling method, meaning that it relies on the laws of probability to select a sample that can be used to make inference to the population; this is the basis of statistical tests of significance .

request a consultation

Discover How We Assist to Edit Your Dissertation Chapters

Aligning theoretical framework, gathering articles, synthesizing gaps, articulating a clear methodology and data plan, and writing about the theoretical and practical implications of your research are part of our comprehensive dissertation editing services.

  • Bring dissertation editing expertise to chapters 1-5 in timely manner.
  • Track all changes, then work with you to bring about scholarly writing.
  • Ongoing support to address committee feedback, reducing revisions.

Random assignment takes place following the selection of participants for the study.  In a true experiment, all study participants are randomly assigned either to receive the treatment (also known as the stimulus or intervention) or to act as a control in the study (meaning they do not receive the treatment).  Although random assignment is a simple procedure (it can be accomplished by the flip of a coin), it can be challenging to implement outside of controlled laboratory conditions.

A study can use both, only one, or neither.  Here are some examples to illustrate each situation:

A researcher gets a list of all students enrolled at a particular school (the population).  Using a random number generator, the researcher selects 100 students from the school to participate in the study (the random sample).  All students’ names are placed in a hat and 50 are chosen to receive the intervention (the treatment group), while the remaining 50 students serve as the control group.  This design uses both random selection and random assignment.

A study using only random assignment could ask the principle of the school to select the students she believes are most likely to enjoy participating in the study, and the researcher could then randomly assign this sample of students to the treatment and control groups.  In such a design the researcher could draw conclusions about the effect of the intervention but couldn’t make any inference about whether the effect would likely to be found in the population.

A study using only random selection could randomly select students from the overall population of the school, but then assign students in one grade to the intervention and students in another grade to the control group.  While any data collected from this sample could be used to make inference to the population of the school, the lack of random assignment to be in the treatment or control group would make it impossible to conclude whether the intervention had any effect.

Random selection is thus essential to external validity, or the extent to which the researcher can use the results of the study to generalize to the larger population.  Random assignment is central to internal validity, which allows the researcher to make causal claims about the effect of the treatment.  Nonrandom assignment often leads to non-equivalent groups, meaning that any effect of the treatment might be a result of the groups being different at the outset rather than different at the end as a result of the treatment.  The consequences of random selection and random assignment are clearly very different, and a strong research design will employ both whenever possible to ensure both internal and external validity .

Popular searches

  • How to Get Participants For Your Study
  • How to Do Segmentation?
  • Conjoint Preference Share Simulator
  • MaxDiff Analysis
  • Likert Scales
  • Reliability & Validity

Request consultation

Do you need support in running a pricing or product study? We can help you with agile consumer research and conjoint analysis.

Looking for an online survey platform?

Conjointly offers a great survey tool with multiple question types, randomisation blocks, and multilingual support. The Basic tier is always free.

Research Methods Knowledge Base

  • Navigating the Knowledge Base
  • Foundations
  • Measurement
  • Internal Validity
  • Introduction to Design
  • Types of Designs
  • Probabilistic Equivalence

Random Selection & Assignment

  • Defining Experimental Designs
  • Factorial Designs
  • Randomized Block Designs
  • Covariance Designs
  • Hybrid Experimental Designs
  • Quasi-Experimental Design
  • Pre-Post Design Relationships
  • Designing Designs for Research
  • Quasi-Experimentation Advances
  • Table of Contents

Fully-functional online survey tool with various question types, logic, randomisation, and reporting for unlimited number of surveys.

Completely free for academics and students .

Random selection is how you draw the sample of people for your study from a population. Random assignment is how you assign the sample that you draw to different groups or treatments in your study.

It is possible to have both random selection and assignment in a study. Let’s say you drew a random sample of 100 clients from a population list of 1000 current clients of your organization. That is random sampling. Now, let’s say you randomly assign 50 of these clients to get some new additional treatment and the other 50 to be controls. That’s random assignment.

It is also possible to have only one of these (random selection or random assignment) but not the other in a study. For instance, if you do not randomly draw the 100 cases from your list of 1000 but instead just take the first 100 on the list, you do not have random selection. But you could still randomly assign this nonrandom sample to treatment versus control. Or, you could randomly select 100 from your list of 1000 and then nonrandomly (haphazardly) assign them to treatment or control.

And, it’s possible to have neither random selection nor random assignment. In a typical nonequivalent groups design in education you might nonrandomly choose two 5th grade classes to be in your study. This is nonrandom selection. Then, you could arbitrarily assign one to get the new educational program and the other to be the control. This is nonrandom (or nonequivalent) assignment.

Random selection is related to sampling . Therefore it is most related to the external validity (or generalizability) of your results. After all, we would randomly sample so that our research participants better represent the larger group from which they’re drawn. Random assignment is most related to design . In fact, when we randomly assign participants to treatments we have, by definition, an experimental design . Therefore, random assignment is most related to internal validity . After all, we randomly assign in order to help assure that our treatment groups are similar to each other (i.e. equivalent) prior to the treatment.

Cookie Consent

Conjointly uses essential cookies to make our site work. We also use additional cookies in order to understand the usage of the site, gather audience analytics, and for remarketing purposes.

For more information on Conjointly's use of cookies, please read our Cookie Policy .

Which one are you?

I am new to conjointly, i am already using conjointly.

Random Assignment in Psychology: Definition & Examples

Julia Simkus

Editor at Simply Psychology

BA (Hons) Psychology, Princeton University

Julia Simkus is a graduate of Princeton University with a Bachelor of Arts in Psychology. She is currently studying for a Master's Degree in Counseling for Mental Health and Wellness in September 2023. Julia's research has been published in peer reviewed journals.

Learn about our Editorial Process

Saul McLeod, PhD

Editor-in-Chief for Simply Psychology

BSc (Hons) Psychology, MRes, PhD, University of Manchester

Saul McLeod, PhD., is a qualified psychology teacher with over 18 years of experience in further and higher education. He has been published in peer-reviewed journals, including the Journal of Clinical Psychology.

Olivia Guy-Evans, MSc

Associate Editor for Simply Psychology

BSc (Hons) Psychology, MSc Psychology of Education

Olivia Guy-Evans is a writer and associate editor for Simply Psychology. She has previously worked in healthcare and educational sectors.

In psychology, random assignment refers to the practice of allocating participants to different experimental groups in a study in a completely unbiased way, ensuring each participant has an equal chance of being assigned to any group.

In experimental research, random assignment, or random placement, organizes participants from your sample into different groups using randomization. 

Random assignment uses chance procedures to ensure that each participant has an equal opportunity of being assigned to either a control or experimental group.

The control group does not receive the treatment in question, whereas the experimental group does receive the treatment.

When using random assignment, neither the researcher nor the participant can choose the group to which the participant is assigned. This ensures that any differences between and within the groups are not systematic at the onset of the study. 

In a study to test the success of a weight-loss program, investigators randomly assigned a pool of participants to one of two groups.

Group A participants participated in the weight-loss program for 10 weeks and took a class where they learned about the benefits of healthy eating and exercise.

Group B participants read a 200-page book that explains the benefits of weight loss. The investigator randomly assigned participants to one of the two groups.

The researchers found that those who participated in the program and took the class were more likely to lose weight than those in the other group that received only the book.

Importance 

Random assignment ensures that each group in the experiment is identical before applying the independent variable.

In experiments , researchers will manipulate an independent variable to assess its effect on a dependent variable, while controlling for other variables. Random assignment increases the likelihood that the treatment groups are the same at the onset of a study.

Thus, any changes that result from the independent variable can be assumed to be a result of the treatment of interest. This is particularly important for eliminating sources of bias and strengthening the internal validity of an experiment.

Random assignment is the best method for inferring a causal relationship between a treatment and an outcome.

Random Selection vs. Random Assignment 

Random selection (also called probability sampling or random sampling) is a way of randomly selecting members of a population to be included in your study.

On the other hand, random assignment is a way of sorting the sample participants into control and treatment groups. 

Random selection ensures that everyone in the population has an equal chance of being selected for the study. Once the pool of participants has been chosen, experimenters use random assignment to assign participants into groups. 

Random assignment is only used in between-subjects experimental designs, while random selection can be used in a variety of study designs.

Random Assignment vs Random Sampling

Random sampling refers to selecting participants from a population so that each individual has an equal chance of being chosen. This method enhances the representativeness of the sample.

Random assignment, on the other hand, is used in experimental designs once participants are selected. It involves allocating these participants to different experimental groups or conditions randomly.

This helps ensure that any differences in results across groups are due to manipulating the independent variable, not preexisting differences among participants.

When to Use Random Assignment

Random assignment is used in experiments with a between-groups or independent measures design.

In these research designs, researchers will manipulate an independent variable to assess its effect on a dependent variable, while controlling for other variables.

There is usually a control group and one or more experimental groups. Random assignment helps ensure that the groups are comparable at the onset of the study.

How to Use Random Assignment

There are a variety of ways to assign participants into study groups randomly. Here are a handful of popular methods: 

  • Random Number Generator : Give each member of the sample a unique number; use a computer program to randomly generate a number from the list for each group.
  • Lottery : Give each member of the sample a unique number. Place all numbers in a hat or bucket and draw numbers at random for each group.
  • Flipping a Coin : Flip a coin for each participant to decide if they will be in the control group or experimental group (this method can only be used when you have just two groups) 
  • Roll a Die : For each number on the list, roll a dice to decide which of the groups they will be in. For example, assume that rolling 1, 2, or 3 places them in a control group and rolling 3, 4, 5 lands them in an experimental group.

When is Random Assignment not used?

  • When it is not ethically permissible: Randomization is only ethical if the researcher has no evidence that one treatment is superior to the other or that one treatment might have harmful side effects. 
  • When answering non-causal questions : If the researcher is just interested in predicting the probability of an event, the causal relationship between the variables is not important and observational designs would be more suitable than random assignment. 
  • When studying the effect of variables that cannot be manipulated: Some risk factors cannot be manipulated and so it would not make any sense to study them in a randomized trial. For example, we cannot randomly assign participants into categories based on age, gender, or genetic factors.

Drawbacks of Random Assignment

While randomization assures an unbiased assignment of participants to groups, it does not guarantee the equality of these groups. There could still be extraneous variables that differ between groups or group differences that arise from chance. Additionally, there is still an element of luck with random assignments.

Thus, researchers can not produce perfectly equal groups for each specific study. Differences between the treatment group and control group might still exist, and the results of a randomized trial may sometimes be wrong, but this is absolutely okay.

Scientific evidence is a long and continuous process, and the groups will tend to be equal in the long run when data is aggregated in a meta-analysis.

Additionally, external validity (i.e., the extent to which the researcher can use the results of the study to generalize to the larger population) is compromised with random assignment.

Random assignment is challenging to implement outside of controlled laboratory conditions and might not represent what would happen in the real world at the population level. 

Random assignment can also be more costly than simple observational studies, where an investigator is just observing events without intervening with the population.

Randomization also can be time-consuming and challenging, especially when participants refuse to receive the assigned treatment or do not adhere to recommendations. 

What is the difference between random sampling and random assignment?

Random sampling refers to randomly selecting a sample of participants from a population. Random assignment refers to randomly assigning participants to treatment groups from the selected sample.

Does random assignment increase internal validity?

Yes, random assignment ensures that there are no systematic differences between the participants in each group, enhancing the study’s internal validity .

Does random assignment reduce sampling error?

Yes, with random assignment, participants have an equal chance of being assigned to either a control group or an experimental group, resulting in a sample that is, in theory, representative of the population.

Random assignment does not completely eliminate sampling error because a sample only approximates the population from which it is drawn. However, random sampling is a way to minimize sampling errors. 

When is random assignment not possible?

Random assignment is not possible when the experimenters cannot control the treatment or independent variable.

For example, if you want to compare how men and women perform on a test, you cannot randomly assign subjects to these groups.

Participants are not randomly assigned to different groups in this study, but instead assigned based on their characteristics.

Does random assignment eliminate confounding variables?

Yes, random assignment eliminates the influence of any confounding variables on the treatment because it distributes them at random among the study groups. Randomization invalidates any relationship between a confounding variable and the treatment.

Why is random assignment of participants to treatment conditions in an experiment used?

Random assignment is used to ensure that all groups are comparable at the start of a study. This allows researchers to conclude that the outcomes of the study can be attributed to the intervention at hand and to rule out alternative explanations for study results.

Further Reading

  • Bogomolnaia, A., & Moulin, H. (2001). A new solution to the random assignment problem .  Journal of Economic theory ,  100 (2), 295-328.
  • Krause, M. S., & Howard, K. I. (2003). What random assignment does and does not do .  Journal of Clinical Psychology ,  59 (7), 751-766.

Print Friendly, PDF & Email

  • Privacy Policy

Research Method

Home » Sampling Methods – Types, Techniques and Examples

Sampling Methods – Types, Techniques and Examples

Table of Contents

Sampling Methods

Sampling refers to the process of selecting a subset of data from a larger population or dataset in order to analyze or make inferences about the whole population.

In other words, sampling involves taking a representative sample of data from a larger group or dataset in order to gain insights or draw conclusions about the entire group.

Sampling Methods

Sampling methods refer to the techniques used to select a subset of individuals or units from a larger population for the purpose of conducting statistical analysis or research.

Sampling is an essential part of the Research because it allows researchers to draw conclusions about a population without having to collect data from every member of that population, which can be time-consuming, expensive, or even impossible.

Types of Sampling Methods

Sampling can be broadly categorized into two main categories:

Probability Sampling

This type of sampling is based on the principles of random selection, and it involves selecting samples in a way that every member of the population has an equal chance of being included in the sample.. Probability sampling is commonly used in scientific research and statistical analysis, as it provides a representative sample that can be generalized to the larger population.

Type of Probability Sampling :

  • Simple Random Sampling: In this method, every member of the population has an equal chance of being selected for the sample. This can be done using a random number generator or by drawing names out of a hat, for example.
  • Systematic Sampling: In this method, the population is first divided into a list or sequence, and then every nth member is selected for the sample. For example, if every 10th person is selected from a list of 100 people, the sample would include 10 people.
  • Stratified Sampling: In this method, the population is divided into subgroups or strata based on certain characteristics, and then a random sample is taken from each stratum. This is often used to ensure that the sample is representative of the population as a whole.
  • Cluster Sampling: In this method, the population is divided into clusters or groups, and then a random sample of clusters is selected. Then, all members of the selected clusters are included in the sample.
  • Multi-Stage Sampling : This method combines two or more sampling techniques. For example, a researcher may use stratified sampling to select clusters, and then use simple random sampling to select members within each cluster.

Non-probability Sampling

This type of sampling does not rely on random selection, and it involves selecting samples in a way that does not give every member of the population an equal chance of being included in the sample. Non-probability sampling is often used in qualitative research, where the aim is not to generalize findings to a larger population, but to gain an in-depth understanding of a particular phenomenon or group. Non-probability sampling methods can be quicker and more cost-effective than probability sampling methods, but they may also be subject to bias and may not be representative of the larger population.

Types of Non-probability Sampling :

  • Convenience Sampling: In this method, participants are chosen based on their availability or willingness to participate. This method is easy and convenient but may not be representative of the population.
  • Purposive Sampling: In this method, participants are selected based on specific criteria, such as their expertise or knowledge on a particular topic. This method is often used in qualitative research, but may not be representative of the population.
  • Snowball Sampling: In this method, participants are recruited through referrals from other participants. This method is often used when the population is hard to reach, but may not be representative of the population.
  • Quota Sampling: In this method, a predetermined number of participants are selected based on specific criteria, such as age or gender. This method is often used in market research, but may not be representative of the population.
  • Volunteer Sampling: In this method, participants volunteer to participate in the study. This method is often used in research where participants are motivated by personal interest or altruism, but may not be representative of the population.

Applications of Sampling Methods

Applications of Sampling Methods from different fields:

  • Psychology : Sampling methods are used in psychology research to study various aspects of human behavior and mental processes. For example, researchers may use stratified sampling to select a sample of participants that is representative of the population based on factors such as age, gender, and ethnicity. Random sampling may also be used to select participants for experimental studies.
  • Sociology : Sampling methods are commonly used in sociological research to study social phenomena and relationships between individuals and groups. For example, researchers may use cluster sampling to select a sample of neighborhoods to study the effects of economic inequality on health outcomes. Stratified sampling may also be used to select a sample of participants that is representative of the population based on factors such as income, education, and occupation.
  • Social sciences: Sampling methods are commonly used in social sciences to study human behavior and attitudes. For example, researchers may use stratified sampling to select a sample of participants that is representative of the population based on factors such as age, gender, and income.
  • Marketing : Sampling methods are used in marketing research to collect data on consumer preferences, behavior, and attitudes. For example, researchers may use random sampling to select a sample of consumers to participate in a survey about a new product.
  • Healthcare : Sampling methods are used in healthcare research to study the prevalence of diseases and risk factors, and to evaluate interventions. For example, researchers may use cluster sampling to select a sample of health clinics to participate in a study of the effectiveness of a new treatment.
  • Environmental science: Sampling methods are used in environmental science to collect data on environmental variables such as water quality, air pollution, and soil composition. For example, researchers may use systematic sampling to collect soil samples at regular intervals across a field.
  • Education : Sampling methods are used in education research to study student learning and achievement. For example, researchers may use stratified sampling to select a sample of schools that is representative of the population based on factors such as demographics and academic performance.

Examples of Sampling Methods

Probability Sampling Methods Examples:

  • Simple random sampling Example : A researcher randomly selects participants from the population using a random number generator or drawing names from a hat.
  • Stratified random sampling Example : A researcher divides the population into subgroups (strata) based on a characteristic of interest (e.g. age or income) and then randomly selects participants from each subgroup.
  • Systematic sampling Example : A researcher selects participants at regular intervals from a list of the population.

Non-probability Sampling Methods Examples:

  • Convenience sampling Example: A researcher selects participants who are conveniently available, such as students in a particular class or visitors to a shopping mall.
  • Purposive sampling Example : A researcher selects participants who meet specific criteria, such as individuals who have been diagnosed with a particular medical condition.
  • Snowball sampling Example : A researcher selects participants who are referred to them by other participants, such as friends or acquaintances.

How to Conduct Sampling Methods

some general steps to conduct sampling methods:

  • Define the population: Identify the population of interest and clearly define its boundaries.
  • Choose the sampling method: Select an appropriate sampling method based on the research question, characteristics of the population, and available resources.
  • Determine the sample size: Determine the desired sample size based on statistical considerations such as margin of error, confidence level, or power analysis.
  • Create a sampling frame: Develop a list of all individuals or elements in the population from which the sample will be drawn. The sampling frame should be comprehensive, accurate, and up-to-date.
  • Select the sample: Use the chosen sampling method to select the sample from the sampling frame. The sample should be selected randomly, or if using a non-random method, every effort should be made to minimize bias and ensure that the sample is representative of the population.
  • Collect data: Once the sample has been selected, collect data from each member of the sample using appropriate research methods (e.g., surveys, interviews, observations).
  • Analyze the data: Analyze the data collected from the sample to draw conclusions about the population of interest.

When to use Sampling Methods

Sampling methods are used in research when it is not feasible or practical to study the entire population of interest. Sampling allows researchers to study a smaller group of individuals, known as a sample, and use the findings from the sample to make inferences about the larger population.

Sampling methods are particularly useful when:

  • The population of interest is too large to study in its entirety.
  • The cost and time required to study the entire population are prohibitive.
  • The population is geographically dispersed or difficult to access.
  • The research question requires specialized or hard-to-find individuals.
  • The data collected is quantitative and statistical analyses are used to draw conclusions.

Purpose of Sampling Methods

The main purpose of sampling methods in research is to obtain a representative sample of individuals or elements from a larger population of interest, in order to make inferences about the population as a whole. By studying a smaller group of individuals, known as a sample, researchers can gather information about the population that would be difficult or impossible to obtain from studying the entire population.

Sampling methods allow researchers to:

  • Study a smaller, more manageable group of individuals, which is typically less time-consuming and less expensive than studying the entire population.
  • Reduce the potential for data collection errors and improve the accuracy of the results by minimizing sampling bias.
  • Make inferences about the larger population with a certain degree of confidence, using statistical analyses of the data collected from the sample.
  • Improve the generalizability and external validity of the findings by ensuring that the sample is representative of the population of interest.

Characteristics of Sampling Methods

Here are some characteristics of sampling methods:

  • Randomness : Probability sampling methods are based on random selection, meaning that every member of the population has an equal chance of being selected. This helps to minimize bias and ensure that the sample is representative of the population.
  • Representativeness : The goal of sampling is to obtain a sample that is representative of the larger population of interest. This means that the sample should reflect the characteristics of the population in terms of key demographic, behavioral, or other relevant variables.
  • Size : The size of the sample should be large enough to provide sufficient statistical power for the research question at hand. The sample size should also be appropriate for the chosen sampling method and the level of precision desired.
  • Efficiency : Sampling methods should be efficient in terms of time, cost, and resources required. The method chosen should be feasible given the available resources and time constraints.
  • Bias : Sampling methods should aim to minimize bias and ensure that the sample is representative of the population of interest. Bias can be introduced through non-random selection or non-response, and can affect the validity and generalizability of the findings.
  • Precision : Sampling methods should be precise in terms of providing estimates of the population parameters of interest. Precision is influenced by sample size, sampling method, and level of variability in the population.
  • Validity : The validity of the sampling method is important for ensuring that the results obtained from the sample are accurate and can be generalized to the population of interest. Validity can be affected by sampling method, sample size, and the representativeness of the sample.

Advantages of Sampling Methods

Sampling methods have several advantages, including:

  • Cost-Effective : Sampling methods are often much cheaper and less time-consuming than studying an entire population. By studying only a small subset of the population, researchers can gather valuable data without incurring the costs associated with studying the entire population.
  • Convenience : Sampling methods are often more convenient than studying an entire population. For example, if a researcher wants to study the eating habits of people in a city, it would be very difficult and time-consuming to study every single person in the city. By using sampling methods, the researcher can obtain data from a smaller subset of people, making the study more feasible.
  • Accuracy: When done correctly, sampling methods can be very accurate. By using appropriate sampling techniques, researchers can obtain a sample that is representative of the entire population. This allows them to make accurate generalizations about the population as a whole based on the data collected from the sample.
  • Time-Saving: Sampling methods can save a lot of time compared to studying the entire population. By studying a smaller sample, researchers can collect data much more quickly than they could if they studied every single person in the population.
  • Less Bias : Sampling methods can reduce bias in a study. If a researcher were to study the entire population, it would be very difficult to eliminate all sources of bias. However, by using appropriate sampling techniques, researchers can reduce bias and obtain a sample that is more representative of the entire population.

Limitations of Sampling Methods

  • Sampling Error : Sampling error is the difference between the sample statistic and the population parameter. It is the result of selecting a sample rather than the entire population. The larger the sample, the lower the sampling error. However, no matter how large the sample size, there will always be some degree of sampling error.
  • Selection Bias: Selection bias occurs when the sample is not representative of the population. This can happen if the sample is not selected randomly or if some groups are underrepresented in the sample. Selection bias can lead to inaccurate conclusions about the population.
  • Non-response Bias : Non-response bias occurs when some members of the sample do not respond to the survey or study. This can result in a biased sample if the non-respondents differ from the respondents in important ways.
  • Time and Cost : While sampling can be cost-effective, it can still be expensive and time-consuming to select a sample that is representative of the population. Depending on the sampling method used, it may take a long time to obtain a sample that is large enough and representative enough to be useful.
  • Limited Information : Sampling can only provide information about the variables that are measured. It may not provide information about other variables that are relevant to the research question but were not measured.
  • Generalization : The extent to which the findings from a sample can be generalized to the population depends on the representativeness of the sample. If the sample is not representative of the population, it may not be possible to generalize the findings to the population as a whole.

About the author

' src=

Muhammad Hassan

Researcher, Academic Writer, Web developer

You may also like

Stratified Sampling

Stratified Random Sampling – Definition, Method...

Purposive Sampling

Purposive Sampling – Methods, Types and Examples

Quota Sampling

Quota Sampling – Types, Methods and Examples

Simple Random Sampling

Simple Random Sampling – Types, Method and...

Non-probability Sampling

Non-probability Sampling – Types, Methods and...

Systematic Sampling

Systematic Sampling – Types, Method and Examples

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • Korean J Anesthesiol
  • v.72(3); 2019 Jun

Randomization in clinical studies

Chi-yeon lim.

1 Department of Biostatistics, Dongguk University College of Medicine, Goyang, Korea

2 Department of Anesthesiology and Pain Medicine, Dongguk University Ilsan Hospital, Goyang, Korea

Randomized controlled trial is widely accepted as the best design for evaluating the efficacy of a new treatment because of the advantages of randomization (random allocation). Randomization eliminates accidental bias, including selection bias, and provides a base for allowing the use of probability theory. Despite its importance, randomization has not been properly understood. This article introduces the different randomization methods with examples: simple randomization; block randomization; adaptive randomization, including minimization; and response-adaptive randomization. Ethics related to randomization are also discussed. The study is helpful in understanding the basic concepts of randomization and how to use R software.

Introduction

Statistical inference in clinical trials is a mandatory process to verify the efficacy and safety of drugs, medical devices, and procedures. It allows for generalizing the results observed through sample, so the sample by random sampling is very important. A randomized controlled trial (RCT) comparing the effects among study groups carry out to avoid any bias at the stage of the planning a study protocol. Randomization (or random allocation of subjects) can mitigate these biases with its randomness, which implies no rule or predictability for allocating subjects to treatment and control groups.

Another property of randomization is that it promotes comparability of the study groups and serves as a basis for statistical inference for quantitative evaluation of the treatment effect. Randomization can be used to create similarity of groups. In other words, all factors, whether known or unknown, that may affect the outcome can be similarly distributed among groups. This similarity is very important and allows for statistical inferences on the treatment effects. Also, it ensures that other factors except treatment do not affect the outcome. If the outcomes of the treatment group and control group show differences, this will be the only difference between the groups, leading to the conclusion that the difference is treatment induced [ 1 ].

CONSORT 1 ) , a set of guidelines proposed to improve completeness of the clinical study report, also includes randomization. Randomization plays a crucial role in increasing the quality of evidence-based studies by minimizing the selection bias that could affect the outcomes. In general, randomization places programming for random number generation, random allocation concealment for security, and a separate random code manager. After then, the generated randomization is implemented to the study [ 2 ]. Randomization is based on probability theory and hence difficult to understand. Moreover, its reproducibility problem requires the use of computer programming language. This study tries to alleviate these difficulties by enabling even a non-statistician to understand randomization for a comparative RCT design.

Methods of Randomization

The method of randomization applied must be determined at the planning stage of a study. “Randomness” cannot be predicted because it involves no rule, constraint, or characteristic. Randomization can minimize the predictability of which treatment will be performed. The method described here is called simple randomization (or complete randomization). However, the absence of rules, constraints, or characteristics does not completely eliminate imbalances by chance. For example, assume that in a multicenter study, all subjects are randomly allocated to treatment or control groups. If subjects from center A are mainly allocated to the control group and lots of subjects from center B are allocated to the treatment group, even though this is allocated with simple randomization, can we ignore the imbalance of the randomization rate in each center?

For another example, if the majority of subjects in the control group were recruited early in the study and/or the majority of those in the treatment group were recruited later in the study, can the chronological bias be ignored? The imbalance in simple randomization is often resolved through restrictive randomization, which is a slightly restricted method [ 3 , 4 ]. Furthermore, adaptive randomization can change the allocation of subjects to reflect the prognostic factors or the response to therapy during the study. The use of adaptive randomization has been increasing in recent times, but simple or restrictive randomization continues to be widely used [ 4 ]. In the Appendix , the R commands are prepared for the various randomization methods described below.

Simple randomization

In simple randomization, a coin or a die roll, for example, may be used to allocate subjects to a group. The best part of simple randomization is that it minimizes any bias by eliminating predictability. Furthermore, each subject can maintain complete randomness and independence with regard to the treatment administered [ 5 ]. This method is easy to understand and apply, 2 ) but it cannot prevent the imbalances in the sample size or prognostic factors that are likely to occur as the number of subjects participating in the study decreases. If the ratio of number of subjects shows an imbalance, that is, it is not 1 : 1, even with the same number of subjects participating, the power of the study will fall. In a study involving a total of 40 subjects in two groups, if 20 subjects are allocated to each group, the power is 80%; this will be 77% for a 25/15 subject allocation and 67% for a 30/10 subject allocation ( Fig. 1 ). 3 ) In addition, it would be difficult to consider a 25/15 or 30/10 subject allocation as aesthetically balanced. 4 ) In other words, the balancing of subjects seems plausible to both researchers and readers. Unfortunately, the nature of simple randomization rarely lets the number of subjects in both groups to be equal [ 6 ]. Therefore, if it is not out of the range of the assignment ratio (e.g., 45%–55%), 5 ) it is balanced. As the total number of subjects increases, the probability of departing from the assignment ratio, that is, the probability of imbalance, decreases. In the following, the total number of subjects and the probability of imbalance were examined in the two-group study with an assignment ratio of 45%–55% ( Fig. 2 ). If the total number of subjects is 40, the probability of the imbalance is 52.7% ( Fig. 2 , point A), but this decreases to 15.7% for 200 subjects ( Fig. 2 , point B) and 4.6% for 400 subjects ( Fig. 2 , point C). This is the randomization method recommended for large-scale clinical trials, because the likelihood of imbalance in trials with a small number of subjects is high [ 6 – 8 ]. 6 ) However, as the number of subjects does not always increase, other solutions need to be considered. A block randomization is helpful to resolve the imbalance in number of subjects, while a stratified randomization and an adaptive randomization can help resolve the imbalance in prognostic factors.

An external file that holds a picture, illustration, etc.
Object name is kja-19049f1.jpg

Influence of sample size ratio in two groups on power (difference (d) = 0.9, two-tailed, significant level = 0.05). The dashed line indicates the same sample size in two groups (n = 20) and maximized power.

An external file that holds a picture, illustration, etc.
Object name is kja-19049f2.jpg

Probability curves of imbalance between two groups for complete randomization as a function of total sample size (n). When n = 40, there is a 52.7% chance of imbalance beyond 10% (allocation ratio 45%–55%) (point A). When n = 200, there is a 15.7% chance of imbalance (point B), but n = 400 results in only 4.6% chance of imbalance (point C).

Block randomization

If we consider only the balance in number of subjects in a study involving two treatment groups A and B, then A and B can be repeatedly allocated in a randomized block design with predefined block size. Here, a selection bias is inevitable because a researcher or subject can easily predict the allocation of the group. For a small number of subjects, their number in the treatment groups will not remain the same as the study progresses, and the statistical analysis may show the problem of poor power. To avoid this, we set blocks for randomization and balance the number of subjects in each block. 7 ) When using blocks, we need to apply multiple blocks and randomize within each block. At the end of block randomization, the number of subjects can easily be balanced, and the maximum imbalance in the study can be limited to an appropriate level. That is, block randomization has the advantage of increasing the comparability between groups by keeping the ratio of the number of subjects between groups almost the same. However, if the block size is 2, the allocation result of the second subject in the block can be easily predicted with a high risk of observation bias. 8 ) Therefore, the block size used should preferably be 4 or more. However, note that even when the block size is large, if the block size is known to the researcher, the risk of selection bias will increase because the treatment of the last subject in the block will be revealed. To reduce the risk of predictability from the use of one block size, the size may be varied. 9 )

Restricted randomization for unbalanced allocation

Sometimes unbalanced allocation becomes necessary for ethical or cost reasons [ 9 ]. Furthermore, if you expect a high dropout rate in a particular group, you have to allocate more subjects. For example, for patients with terminal cancer who are not treated with conventional anticancer agents, it would be both ethical and helpful to recruit those who would be more likely to receive a newly developed anticancer drug [ 10 ] (of course, contrary to expectations, the drug could be harmful).

As for simple randomization, the probability is first determined according to the ratio between the groups, and then the subjects are allocated. If the ratio between group A and group B is 2 : 1, the probability of group A is 2/3 and that of group B is 1/3. Block randomization often uses a jar model with a random allocation rule. To consider the method, first drop as many balls as the number of subjects into the jar according to the group allocation ratio (of course, the balls have different colors depending on the group). Whenever you allocate a subject, take out one ball randomly and confirm it, and do not place the ball back into the jar (random sampling without replacement). Repeat this allocation for each block.

Stratified randomization

Some studies have prognostic factors or covariates affecting the study outcome as well as treatment. Researchers hope to balance the prognostic factors between the study groups, but randomization does not eliminate all the imbalances in prognostic factors. Stratified randomization refers to the situation where the strata are based on level of prognostic factors or covariates. For example, if “sex” is the chosen prognostic factor, the number of strata is two (male and female), and randomization is applied to each stratum. When a male subject participates, the subject is first allocated to the male strata, and the group (treatment group, control group, etc.) is determined through randomization applied to the male strata. In a multicenter study, one typical prognostic factor is the “site.” This may be due to the differences in characteristics between the subjects and the manner and procedure in which the patients are treated in each hospital.

Stratification can reduce imbalances and increase statistical power, but it has certain problems. If several important prognostic factors affect the outcome, the number of strata would increase [ 11 ]. For example, 12 (2 × 2 × 3) strata are formed solely from recruitment hospitals (sites 1 and 2), sex (male and female), and age group (under 20 years, 20–64 years, and 65 years and older) ( Fig. 3 ). In case of several strata in relation to the target sample size, the number of subjects allocated to a few strata may be empty or sparse. This causes an imbalance10) 10 ) in the number of subjects allocated to the treatment group. To reduce this risk, the prognostic factors should be carefully selected. These prognostic factors should be considered again during the statistical analysis and at the end of the study.

An external file that holds a picture, illustration, etc.
Object name is kja-19049f3.jpg

Example of stratification with three prognostic factors (site, sex, and age band). Eventually, randomization with 12 strata should be accomplished using 12 separate randomization processes. C: control group, T: treatment group.

Adaptive randomization

Adaptive randomization is a method of changing the allocation probability according to the progress and position of the study. It may be used to minimize the imbalance between treatment groups as well as to change the allocation probability based on the therapeutic effect. Covariate-adaptive randomization adjusts the allocation of each subject to reduce the imbalance, taking into account the imbalance of the prognostic factors. One example is the “minimization technique of randomization (minimization)” to develop indicators that collectively determine the distributional imbalance of various prognostic factors and allocates them to minimize the imbalance.

Minimization 11 )

Minimization was first introduced as a covariate adaptive method to balance the prognostic factors [ 12 , 13 ]. The first subject is allocated through simple randomization, and the subsequent ones are allocated to balance the prognostic factors. In other words, the information of the subjects who have already participated in the study is used to allocate the newly recruited subjects and minimize the imbalance of the prognostic factors [ 14 ].

Several methods have emerged following Taves [ 13 ]. Pocock and Simon define a more general method [ 12 ]. 12 ) First, the total number of imbalances is calculated after virtually allocating a newly recruited subject to all groups, respectively. Then, each group has its own the total number of imbalances. Here, this subject will be allocated to the group with lowest total number of imbalances.

We next proceed with a virtual allocation to the recruitment hospitals (Sites 1 and 2), sex (male and female), and age band (under 20 years, 20–64 years, and 65 years or older) as prognostic factors. This study has two groups: a treatment group and a control group.

Assume that the first subject (male, 52-years-old) was recruited from Site 2. Because this subject is the first one, the allocation is determined by simple randomization.

Further, assume that the subject is allocated to a treatment group. In this group, scores are added to Site 2 of the recruiting hospital, sex Male, and the 20–64 age band ( Table 1 ). Next, assume that the second subject (female, 25-years-old) was recruited through Site 2. Calculate the total number of imbalances when this subject is allocated to the treatment group and to the control group. Add the appropriate scores to the area within each group, and sum the differences between the areas.

How Adaptive Randomization Using Minimization Works

Prognostic factorControl groupTreatment group
Site
 Site 100
 Site 201
Sex
 Male01
 Female00
Age band
 < 2000
 20–6401
 ≥ 6500

The score in each factor is 0. The first patient (sex male, 52 yr, from site 2) is allocated to the treatment group through simple randomization. Therefore, site 2, sex male, and the 20–64 years age band in the treatment group receive the score.

First, the total number of imbalances when the subject is allocated to the control group is

The total number of imbalances when the subject is allocated to the treatment group is

Since the total number of imbalances when the subject is allocated to the control group has 1 point (< 5), the second subject is allocated to the control group, and the score is added to Site 2 of the recruiting hospital, Sex female, and the 20–64 age band in the control group ( Table 2 ). Next, the third subject (Site 1, Sex male, 17-years-old) is recruited.

Prognostic factorControl groupTreatment group
If allocated to control groupSite
 Site 100
 Site 211
Sex
 Male01
 Female10
Age band
 < 2000
 20–6411
 ≥ 6500
Total number of imbalances[(1 − 1) + (1 − 0) + (1 − 1)] = 1
If allocated to treatment groupSite
 Site 100
 Site 202
Sex
 Male01
 Female01
Age band
 < 2000
 20–6402
 ≥ 6500
Total number of imbalances[(2 − 0) + (1 − 0) + (2 − 0)] = 5

The second patient has factors sex female, 25 yr, and site 2. If this patient is allocated to the control group, the total imbalance is 1. If this patient is allocated to the treatment group, the total imbalance is 5. Therefore, this patient is allocated to the control group, and site 2, sex female, and the 20–64 years age band in the control group receive the score.

Now, the total number of imbalances when the subject is allocated to the control group is

The total number of imbalances when the subject is allocated to the control group is 2 point (< 4). Therefore, the third subject is allocated to the control group, and the score is added to Site 1 of the recruiting hospital, sex male, and the < 20 age band ( Table 3 ). The subjects are allocated and scores added in this manner. Now, assume that the study continues, and the 15th subject (female, 74-years-old) is recruited from Site 2.

Prognostic factorControl groupTreatment group
If allocated to control groupSite
 Site 110
 Site 211
Sex
 Male11
 Female10
Age band
 < 2010
 20–6411
 ≥ 6500
Total number of imbalances[(1 − 0) + (1 − 1) + (1 − 0)] = 2
If allocated to treatment groupSite
 Site 101
 Site 211
Sex
 Male02
 Female10
Age band
 < 2001
 20–6411
 ≥ 6500
Total number of imbalances[(1 − 0) + (2 − 0) + (1 − 0)] = 4

The third patient has factors sex male, 17 yr, and site 1. If this patient is allocated to the control group, the total imbalance is 2. If this patient is allocated to the treatment group, the total imbalance is 4. Therefore,this patient is allocated to the control group, and then site 1, sex male, and the < 20 age band in the control group receive the score.

Here, the total number of imbalances when the subject is allocated to the control group is

The total number of imbalances when the subject is allocated to the control group is lower than that when the allocation is to the treatment group (3 < 5). Therefore, the 15th subject is allocated to the control group, and the score is added to Site 2 of the recruiting hospital, female sex, and the ≥ 65 age band ( Table 4 ). If the total number of imbalances during the minimization technique is the same, the allocation is determined by simple randomization.

Prognostic factorControl groupTreatment group
If allocated to control groupSite
 Site 142
 Site 245
Sex
Male44
 Female43
Age band
 < 2022
 20–6422
 ≥ 6543
Total number of imbalances[(5 − 4) + (4 − 3) + (4 − 3)] = 3
If allocated to treatment groupSite
 Site 142
 Site 236
Sex
 Male44
 Female34
Age band
 < 2022
 20–6422
 ≥ 6534
Total number of imbalances[(6 − 3) + (4 − 3) + (4 − 3)] = 5

The 15th patient has factors sex female, 74 yr, and site 2. If this patient is allocated to the control group, the total imbalance is 3. If this patient is allocated to the treatment group, the total imbalance is 5. Therefore, this patient is allocated to the control group, and site 2, sex female, and the ≥ 65 age band in the control group receive the score.

Although minimization is designed to overcome the disadvantages of stratified randomization, this method also has drawbacks. A concern from a statistical point of view is that it does not satisfy randomness, which is the basic assumption of statistical inference [ 15 , 16 ]. For this reason, the analysis of covariance or permutation test are proposed [ 13 ]. Furthermore, exposure of the subjects’ information can lead to a certain degree of allocation prediction for the next subjects. The calculation process is complicated, but can be carried out through various programs.

Response-adaptive randomization

So far, the randomization methods is assumed that the variances of treatment effects are equal in each group. Thus, the number of subjects in both groups is determined under this assumption. However, when analyzing the data accruing as the study progresses, what happens if the variance in treatment effects is not the same? In this case, would it not reduce the number of subjects initially determined rather than the statistical power? In other words, should the allocation probabilities determined prior to the study remain constant throughout the study? Alternatively, is it possible to change the allocation probability during the study by using the data accruing as the study progresses? If the treatment effects turn out to be inferior during the study, would it be advisable to reduce the number of subjects allocated to this group [ 17 , 18 ]?

An example of response-adaptive randomization is the randomized play-the-winner rule. Here, the first subject is allocated by predefined randomization, and if this patient’s response is “success,” the next patient will be allocated to the same treatment group; otherwise, the patient will be allocated to another treatment. That is, this method is based on statistical reasoning that is not possible under a fixed allocation probability and on the ethics of allowing more patients to be allocated to treatments that benefit the patients. However, the method can lead to imbalances between the treatment groups. In addition, if clinical studies take a very long time to obtain the results of patient responses, this method cannot be recommended.

Ethics of Randomization

As noted earlier, RCT is a scientific study design based on the probability of allocating subjects to treatment groups in order to ensure comparability, form the basis of statistical inference, and identify the effects of treatment. However, an ethical debate needs to examine whether the treatment method for the subjects, especially for patients, should be determined by probability rather than by the physician. Nonetheless, the decisions should preferably be made by probability because clinical trials have the distinct goals of investigating the efficacy and safety of new medicines, medical devices, and procedures, rather than merely reach therapeutic conclusions. The purpose of the study is therefore to maintain objectivity, which is why prejudice and bias should be excluded. That is, only an unconstrained attitude during the study can confirm that a particular medicine, medical device, or procedure is effective or safe.

Consider this from another perspective. If the researcher maintains an unconstrained attitude, and the subject receives all the information, understands it, and decides to voluntarily participate, is the clinical study ethical? Unfortunately, this is not so easy to answer. Participation in a clinical study may provide the subject with the benefit of treatment, but it could be risky. Furthermore, the subjects may be given a placebo, and not treatment. Eventually, the subject may be forced to make personal sacrifices for ambiguous benefit. In other words, some subjects have to undergo further treatment, representing the cost that society has to pay for the benefit of future subjects or for a larger number of subjects [ 4 , 19 ]. This ethical dilemma on the balance between individual ethics and collective ethics [ 20 ] is still spawning much controversy. If, additionally, the researcher is biased, the controversy over this dilemma will obviously become more confused and the reliability of the study will be lowered. Therefore, randomization is a key factor in a study having to clarify causality through comparison.

Conclusions

Studies have described a random table with subsequent randomization. However, if accurate information on randomization is not provided, it would be difficult to gain enough confidence to proceed with the study and arrive at conclusions. Furthermore, probability-based treatment is allowed with the hope that the trial will be conducted through proper processes, and that the outcome will ultimately benefit the medical profession. Concurrently, it should be fully appreciated that the contribution of the subjects involved in this process is a social cost.

1) http://www.consort-statement.org/

2) However, since the results and process of randomization cannot be easily recorded, the audit of randomization is difficult.

3) Two-tailed test with difference (d) = 0.91 and type 1 error of 0.05.

4) “Cosmetic credibility” is often used.

5) The difference in number of subjects does not exceed 10% of the total number of subjects. This range is determined by a researcher, who is also able to choose 20% instead of 10%.

6) These references recommend 200 or more subjects, but it is not possible to determine the exact number.

7) Random allocation rule, truncated binomial randomization, Hadamard randomization, and the maximal procedure are forced balance randomization methods within blocks, and one of them is applied to the block. The details are beyond the scope of this study, and are therefore not covered.

8) The block size of 2 applies mainly to a study of allocating a pair at the same time.

9) Strictly speaking, the block size is randomly selected from a discrete uniform distribution, and so the use of a random block design rather than a “varying” block size would be a more formal procedure.

10) As the number of strata increases, the imbalance increases due to various factors. The details are beyond the scope of this study.

11) This paragraph introduces how to allocate “two” groups.

12) We can set the weights on the variables or the allowable range for the total number of imbalance, but in this study, we did not set any weights or allowable range for the total number of imbalances.

No potential conflict of interest relevant to this article was reported.

Authors’ contribution

Chi-Yeon Lim (Software; Supervision; Validation; Writing – review & editing)

Junyong In (Conceptualization; Software; Visualization; Writing – original draft; Writing – review & editing)

Frequently asked questions

What’s the difference between random assignment and random selection.

Random selection, or random sampling , is a way of selecting members of a population for your study’s sample.

In contrast, random assignment is a way of sorting the sample into control and experimental groups.

Random sampling enhances the external validity or generalizability of your results, while random assignment improves the internal validity of your study.

Frequently asked questions: Methodology

Attrition refers to participants leaving a study. It always happens to some extent—for example, in randomized controlled trials for medical research.

Differential attrition occurs when attrition or dropout rates differ systematically between the intervention and the control group . As a result, the characteristics of the participants who drop out differ from the characteristics of those who stay in the study. Because of this, study results may be biased .

Action research is conducted in order to solve a particular issue immediately, while case studies are often conducted over a longer period of time and focus more on observing and analyzing a particular ongoing phenomenon.

Action research is focused on solving a problem or informing individual and community-based knowledge in a way that impacts teaching, learning, and other related processes. It is less focused on contributing theoretical input, instead producing actionable input.

Action research is particularly popular with educators as a form of systematic inquiry because it prioritizes reflection and bridges the gap between theory and practice. Educators are able to simultaneously investigate an issue as they solve it, and the method is very iterative and flexible.

A cycle of inquiry is another name for action research . It is usually visualized in a spiral shape following a series of steps, such as “planning → acting → observing → reflecting.”

To make quantitative observations , you need to use instruments that are capable of measuring the quantity you want to observe. For example, you might use a ruler to measure the length of an object or a thermometer to measure its temperature.

Criterion validity and construct validity are both types of measurement validity . In other words, they both show you how accurately a method measures something.

While construct validity is the degree to which a test or other measurement method measures what it claims to measure, criterion validity is the degree to which a test can predictively (in the future) or concurrently (in the present) measure something.

Construct validity is often considered the overarching type of measurement validity . You need to have face validity , content validity , and criterion validity in order to achieve construct validity.

Convergent validity and discriminant validity are both subtypes of construct validity . Together, they help you evaluate whether a test measures the concept it was designed to measure.

  • Convergent validity indicates whether a test that is designed to measure a particular construct correlates with other tests that assess the same or similar construct.
  • Discriminant validity indicates whether two tests that should not be highly related to each other are indeed not related. This type of validity is also called divergent validity .

You need to assess both in order to demonstrate construct validity. Neither one alone is sufficient for establishing construct validity.

  • Discriminant validity indicates whether two tests that should not be highly related to each other are indeed not related

Content validity shows you how accurately a test or other measurement method taps  into the various aspects of the specific construct you are researching.

In other words, it helps you answer the question: “does the test measure all aspects of the construct I want to measure?” If it does, then the test has high content validity.

The higher the content validity, the more accurate the measurement of the construct.

If the test fails to include parts of the construct, or irrelevant parts are included, the validity of the instrument is threatened, which brings your results into question.

Face validity and content validity are similar in that they both evaluate how suitable the content of a test is. The difference is that face validity is subjective, and assesses content at surface level.

When a test has strong face validity, anyone would agree that the test’s questions appear to measure what they are intended to measure.

For example, looking at a 4th grade math test consisting of problems in which students have to add and multiply, most people would agree that it has strong face validity (i.e., it looks like a math test).

On the other hand, content validity evaluates how well a test represents all the aspects of a topic. Assessing content validity is more systematic and relies on expert evaluation. of each question, analyzing whether each one covers the aspects that the test was designed to cover.

A 4th grade math test would have high content validity if it covered all the skills taught in that grade. Experts(in this case, math teachers), would have to evaluate the content validity by comparing the test to the learning objectives.

Snowball sampling is a non-probability sampling method . Unlike probability sampling (which involves some form of random selection ), the initial individuals selected to be studied are the ones who recruit new participants.

Because not every member of the target population has an equal chance of being recruited into the sample, selection in snowball sampling is non-random.

Snowball sampling is a non-probability sampling method , where there is not an equal chance for every member of the population to be included in the sample .

This means that you cannot use inferential statistics and make generalizations —often the goal of quantitative research . As such, a snowball sample is not representative of the target population and is usually a better fit for qualitative research .

Snowball sampling relies on the use of referrals. Here, the researcher recruits one or more initial participants, who then recruit the next ones.

Participants share similar characteristics and/or know each other. Because of this, not every member of the population has an equal chance of being included in the sample, giving rise to sampling bias .

Snowball sampling is best used in the following cases:

  • If there is no sampling frame available (e.g., people with a rare disease)
  • If the population of interest is hard to access or locate (e.g., people experiencing homelessness)
  • If the research focuses on a sensitive topic (e.g., extramarital affairs)

The reproducibility and replicability of a study can be ensured by writing a transparent, detailed method section and using clear, unambiguous language.

Reproducibility and replicability are related terms.

  • Reproducing research entails reanalyzing the existing data in the same manner.
  • Replicating (or repeating ) the research entails reconducting the entire analysis, including the collection of new data . 
  • A successful reproduction shows that the data analyses were conducted in a fair and honest manner.
  • A successful replication shows that the reliability of the results is high.

Stratified sampling and quota sampling both involve dividing the population into subgroups and selecting units from each subgroup. The purpose in both cases is to select a representative sample and/or to allow comparisons between subgroups.

The main difference is that in stratified sampling, you draw a random sample from each subgroup ( probability sampling ). In quota sampling you select a predetermined number or proportion of units, in a non-random manner ( non-probability sampling ).

Purposive and convenience sampling are both sampling methods that are typically used in qualitative data collection.

A convenience sample is drawn from a source that is conveniently accessible to the researcher. Convenience sampling does not distinguish characteristics among the participants. On the other hand, purposive sampling focuses on selecting participants possessing characteristics associated with the research study.

The findings of studies based on either convenience or purposive sampling can only be generalized to the (sub)population from which the sample is drawn, and not to the entire population.

Random sampling or probability sampling is based on random selection. This means that each unit has an equal chance (i.e., equal probability) of being included in the sample.

On the other hand, convenience sampling involves stopping people at random, which means that not everyone has an equal chance of being selected depending on the place, time, or day you are collecting your data.

Convenience sampling and quota sampling are both non-probability sampling methods. They both use non-random criteria like availability, geographical proximity, or expert knowledge to recruit study participants.

However, in convenience sampling, you continue to sample units or cases until you reach the required sample size.

In quota sampling, you first need to divide your population of interest into subgroups (strata) and estimate their proportions (quota) in the population. Then you can start your data collection, using convenience sampling to recruit participants, until the proportions in each subgroup coincide with the estimated proportions in the population.

A sampling frame is a list of every member in the entire population . It is important that the sampling frame is as complete as possible, so that your sample accurately reflects your population.

Stratified and cluster sampling may look similar, but bear in mind that groups created in cluster sampling are heterogeneous , so the individual characteristics in the cluster vary. In contrast, groups created in stratified sampling are homogeneous , as units share characteristics.

Relatedly, in cluster sampling you randomly select entire groups and include all units of each group in your sample. However, in stratified sampling, you select some units of all groups and include them in your sample. In this way, both methods can ensure that your sample is representative of the target population .

A systematic review is secondary research because it uses existing research. You don’t collect new data yourself.

The key difference between observational studies and experimental designs is that a well-done observational study does not influence the responses of participants, while experiments do have some sort of treatment condition applied to at least some participants by random assignment .

An observational study is a great choice for you if your research question is based purely on observations. If there are ethical, logistical, or practical concerns that prevent you from conducting a traditional experiment , an observational study may be a good choice. In an observational study, there is no interference or manipulation of the research subjects, as well as no control or treatment groups .

It’s often best to ask a variety of people to review your measurements. You can ask experts, such as other researchers, or laypeople, such as potential participants, to judge the face validity of tests.

While experts have a deep understanding of research methods , the people you’re studying can provide you with valuable insights you may have missed otherwise.

Face validity is important because it’s a simple first step to measuring the overall validity of a test or technique. It’s a relatively intuitive, quick, and easy way to start checking whether a new measure seems useful at first glance.

Good face validity means that anyone who reviews your measure says that it seems to be measuring what it’s supposed to. With poor face validity, someone reviewing your measure may be left confused about what you’re measuring and why you’re using this method.

Face validity is about whether a test appears to measure what it’s supposed to measure. This type of validity is concerned with whether a measure seems relevant and appropriate for what it’s assessing only on the surface.

Statistical analyses are often applied to test validity with data from your measures. You test convergent validity and discriminant validity with correlations to see if results from your test are positively or negatively related to those of other established tests.

You can also use regression analyses to assess whether your measure is actually predictive of outcomes that you expect it to predict theoretically. A regression analysis that supports your expectations strengthens your claim of construct validity .

When designing or evaluating a measure, construct validity helps you ensure you’re actually measuring the construct you’re interested in. If you don’t have construct validity, you may inadvertently measure unrelated or distinct constructs and lose precision in your research.

Construct validity is often considered the overarching type of measurement validity ,  because it covers all of the other types. You need to have face validity , content validity , and criterion validity to achieve construct validity.

Construct validity is about how well a test measures the concept it was designed to evaluate. It’s one of four types of measurement validity , which includes construct validity, face validity , and criterion validity.

There are two subtypes of construct validity.

  • Convergent validity : The extent to which your measure corresponds to measures of related constructs
  • Discriminant validity : The extent to which your measure is unrelated or negatively related to measures of distinct constructs

Naturalistic observation is a valuable tool because of its flexibility, external validity , and suitability for topics that can’t be studied in a lab setting.

The downsides of naturalistic observation include its lack of scientific control , ethical considerations , and potential for bias from observers and subjects.

Naturalistic observation is a qualitative research method where you record the behaviors of your research subjects in real world settings. You avoid interfering or influencing anything in a naturalistic observation.

You can think of naturalistic observation as “people watching” with a purpose.

A dependent variable is what changes as a result of the independent variable manipulation in experiments . It’s what you’re interested in measuring, and it “depends” on your independent variable.

In statistics, dependent variables are also called:

  • Response variables (they respond to a change in another variable)
  • Outcome variables (they represent the outcome you want to measure)
  • Left-hand-side variables (they appear on the left-hand side of a regression equation)

An independent variable is the variable you manipulate, control, or vary in an experimental study to explore its effects. It’s called “independent” because it’s not influenced by any other variables in the study.

Independent variables are also called:

  • Explanatory variables (they explain an event or outcome)
  • Predictor variables (they can be used to predict the value of a dependent variable)
  • Right-hand-side variables (they appear on the right-hand side of a regression equation).

As a rule of thumb, questions related to thoughts, beliefs, and feelings work well in focus groups. Take your time formulating strong questions, paying special attention to phrasing. Be careful to avoid leading questions , which can bias your responses.

Overall, your focus group questions should be:

  • Open-ended and flexible
  • Impossible to answer with “yes” or “no” (questions that start with “why” or “how” are often best)
  • Unambiguous, getting straight to the point while still stimulating discussion
  • Unbiased and neutral

A structured interview is a data collection method that relies on asking questions in a set order to collect data on a topic. They are often quantitative in nature. Structured interviews are best used when: 

  • You already have a very clear understanding of your topic. Perhaps significant research has already been conducted, or you have done some prior research yourself, but you already possess a baseline for designing strong structured questions.
  • You are constrained in terms of time or resources and need to analyze your data quickly and efficiently.
  • Your research question depends on strong parity between participants, with environmental conditions held constant.

More flexible interview options include semi-structured interviews , unstructured interviews , and focus groups .

Social desirability bias is the tendency for interview participants to give responses that will be viewed favorably by the interviewer or other participants. It occurs in all types of interviews and surveys , but is most common in semi-structured interviews , unstructured interviews , and focus groups .

Social desirability bias can be mitigated by ensuring participants feel at ease and comfortable sharing their views. Make sure to pay attention to your own body language and any physical or verbal cues, such as nodding or widening your eyes.

This type of bias can also occur in observations if the participants know they’re being observed. They might alter their behavior accordingly.

The interviewer effect is a type of bias that emerges when a characteristic of an interviewer (race, age, gender identity, etc.) influences the responses given by the interviewee.

There is a risk of an interviewer effect in all types of interviews , but it can be mitigated by writing really high-quality interview questions.

A semi-structured interview is a blend of structured and unstructured types of interviews. Semi-structured interviews are best used when:

  • You have prior interview experience. Spontaneous questions are deceptively challenging, and it’s easy to accidentally ask a leading question or make a participant uncomfortable.
  • Your research question is exploratory in nature. Participant answers can guide future research questions and help you develop a more robust knowledge base for future research.

An unstructured interview is the most flexible type of interview, but it is not always the best fit for your research topic.

Unstructured interviews are best used when:

  • You are an experienced interviewer and have a very strong background in your research topic, since it is challenging to ask spontaneous, colloquial questions.
  • Your research question is exploratory in nature. While you may have developed hypotheses, you are open to discovering new or shifting viewpoints through the interview process.
  • You are seeking descriptive data, and are ready to ask questions that will deepen and contextualize your initial thoughts and hypotheses.
  • Your research depends on forming connections with your participants and making them feel comfortable revealing deeper emotions, lived experiences, or thoughts.

The four most common types of interviews are:

  • Structured interviews : The questions are predetermined in both topic and order. 
  • Semi-structured interviews : A few questions are predetermined, but other questions aren’t planned.
  • Unstructured interviews : None of the questions are predetermined.
  • Focus group interviews : The questions are presented to a group instead of one individual.

Deductive reasoning is commonly used in scientific research, and it’s especially associated with quantitative research .

In research, you might have come across something called the hypothetico-deductive method . It’s the scientific method of testing hypotheses to check whether your predictions are substantiated by real-world data.

Deductive reasoning is a logical approach where you progress from general ideas to specific conclusions. It’s often contrasted with inductive reasoning , where you start with specific observations and form general conclusions.

Deductive reasoning is also called deductive logic.

There are many different types of inductive reasoning that people use formally or informally.

Here are a few common types:

  • Inductive generalization : You use observations about a sample to come to a conclusion about the population it came from.
  • Statistical generalization: You use specific numbers about samples to make statements about populations.
  • Causal reasoning: You make cause-and-effect links between different things.
  • Sign reasoning: You make a conclusion about a correlational relationship between different things.
  • Analogical reasoning: You make a conclusion about something based on its similarities to something else.

Inductive reasoning is a bottom-up approach, while deductive reasoning is top-down.

Inductive reasoning takes you from the specific to the general, while in deductive reasoning, you make inferences by going from general premises to specific conclusions.

In inductive research , you start by making observations or gathering data. Then, you take a broad scan of your data and search for patterns. Finally, you make general conclusions that you might incorporate into theories.

Inductive reasoning is a method of drawing conclusions by going from the specific to the general. It’s usually contrasted with deductive reasoning, where you proceed from general information to specific conclusions.

Inductive reasoning is also called inductive logic or bottom-up reasoning.

A hypothesis states your predictions about what your research will find. It is a tentative answer to your research question that has not yet been tested. For some research projects, you might have to write several hypotheses that address different aspects of your research question.

A hypothesis is not just a guess — it should be based on existing theories and knowledge. It also has to be testable, which means you can support or refute it through scientific research methods (such as experiments, observations and statistical analysis of data).

Triangulation can help:

  • Reduce research bias that comes from using a single method, theory, or investigator
  • Enhance validity by approaching the same topic with different tools
  • Establish credibility by giving you a complete picture of the research problem

But triangulation can also pose problems:

  • It’s time-consuming and labor-intensive, often involving an interdisciplinary team.
  • Your results may be inconsistent or even contradictory.

There are four main types of triangulation :

  • Data triangulation : Using data from different times, spaces, and people
  • Investigator triangulation : Involving multiple researchers in collecting or analyzing data
  • Theory triangulation : Using varying theoretical perspectives in your research
  • Methodological triangulation : Using different methodologies to approach the same topic

Many academic fields use peer review , largely to determine whether a manuscript is suitable for publication. Peer review enhances the credibility of the published manuscript.

However, peer review is also common in non-academic settings. The United Nations, the European Union, and many individual nations use peer review to evaluate grant applications. It is also widely used in medical and health-related fields as a teaching or quality-of-care measure. 

Peer assessment is often used in the classroom as a pedagogical tool. Both receiving feedback and providing it are thought to enhance the learning process, helping students think critically and collaboratively.

Peer review can stop obviously problematic, falsified, or otherwise untrustworthy research from being published. It also represents an excellent opportunity to get feedback from renowned experts in your field. It acts as a first defense, helping you ensure your argument is clear and that there are no gaps, vague terms, or unanswered questions for readers who weren’t involved in the research process.

Peer-reviewed articles are considered a highly credible source due to this stringent process they go through before publication.

In general, the peer review process follows the following steps: 

  • First, the author submits the manuscript to the editor.
  • Reject the manuscript and send it back to author, or 
  • Send it onward to the selected peer reviewer(s) 
  • Next, the peer review process occurs. The reviewer provides feedback, addressing any major or minor issues with the manuscript, and gives their advice regarding what edits should be made. 
  • Lastly, the edited manuscript is sent back to the author. They input the edits, and resubmit it to the editor for publication.

Exploratory research is often used when the issue you’re studying is new or when the data collection process is challenging for some reason.

You can use exploratory research if you have a general idea or a specific question that you want to study but there is no preexisting knowledge or paradigm with which to study it.

Exploratory research is a methodology approach that explores research questions that have not previously been studied in depth. It is often used when the issue you’re studying is new, or the data collection process is challenging in some way.

Explanatory research is used to investigate how or why a phenomenon occurs. Therefore, this type of research is often one of the first stages in the research process , serving as a jumping-off point for future research.

Exploratory research aims to explore the main aspects of an under-researched problem, while explanatory research aims to explain the causes and consequences of a well-defined problem.

Explanatory research is a research method used to investigate how or why something occurs when only a small amount of information is available pertaining to that topic. It can help you increase your understanding of a given topic.

Clean data are valid, accurate, complete, consistent, unique, and uniform. Dirty data include inconsistencies and errors.

Dirty data can come from any part of the research process, including poor research design , inappropriate measurement materials, or flawed data entry.

Data cleaning takes place between data collection and data analyses. But you can use some methods even before collecting data.

For clean data, you should start by designing measures that collect valid data. Data validation at the time of data entry or collection helps you minimize the amount of data cleaning you’ll need to do.

After data collection, you can use data standardization and data transformation to clean your data. You’ll also deal with any missing values, outliers, and duplicate values.

Every dataset requires different techniques to clean dirty data , but you need to address these issues in a systematic way. You focus on finding and resolving data points that don’t agree or fit with the rest of your dataset.

These data might be missing values, outliers, duplicate values, incorrectly formatted, or irrelevant. You’ll start with screening and diagnosing your data. Then, you’ll often standardize and accept or remove data to make your dataset consistent and valid.

Data cleaning is necessary for valid and appropriate analyses. Dirty data contain inconsistencies or errors , but cleaning your data helps you minimize or resolve these.

Without data cleaning, you could end up with a Type I or II error in your conclusion. These types of erroneous conclusions can be practically significant with important consequences, because they lead to misplaced investments or missed opportunities.

Data cleaning involves spotting and resolving potential data inconsistencies or errors to improve your data quality. An error is any value (e.g., recorded weight) that doesn’t reflect the true value (e.g., actual weight) of something that’s being measured.

In this process, you review, analyze, detect, modify, or remove “dirty” data to make your dataset “clean.” Data cleaning is also called data cleansing or data scrubbing.

Research misconduct means making up or falsifying data, manipulating data analyses, or misrepresenting results in research reports. It’s a form of academic fraud.

These actions are committed intentionally and can have serious consequences; research misconduct is not a simple mistake or a point of disagreement but a serious ethical failure.

Anonymity means you don’t know who the participants are, while confidentiality means you know who they are but remove identifying information from your research report. Both are important ethical considerations .

You can only guarantee anonymity by not collecting any personally identifying information—for example, names, phone numbers, email addresses, IP addresses, physical characteristics, photos, or videos.

You can keep data confidential by using aggregate information in your research report, so that you only refer to groups of participants rather than individuals.

Research ethics matter for scientific integrity, human rights and dignity, and collaboration between science and society. These principles make sure that participation in studies is voluntary, informed, and safe.

Ethical considerations in research are a set of principles that guide your research designs and practices. These principles include voluntary participation, informed consent, anonymity, confidentiality, potential for harm, and results communication.

Scientists and researchers must always adhere to a certain code of conduct when collecting data from others .

These considerations protect the rights of research participants, enhance research validity , and maintain scientific integrity.

In multistage sampling , you can use probability or non-probability sampling methods .

For a probability sample, you have to conduct probability sampling at every stage.

You can mix it up by using simple random sampling , systematic sampling , or stratified sampling to select units at different stages, depending on what is applicable and relevant to your study.

Multistage sampling can simplify data collection when you have large, geographically spread samples, and you can obtain a probability sample without a complete sampling frame.

But multistage sampling may not lead to a representative sample, and larger samples are needed for multistage samples to achieve the statistical properties of simple random samples .

These are four of the most common mixed methods designs :

  • Convergent parallel: Quantitative and qualitative data are collected at the same time and analyzed separately. After both analyses are complete, compare your results to draw overall conclusions. 
  • Embedded: Quantitative and qualitative data are collected at the same time, but within a larger quantitative or qualitative design. One type of data is secondary to the other.
  • Explanatory sequential: Quantitative data is collected and analyzed first, followed by qualitative data. You can use this design if you think your qualitative data will explain and contextualize your quantitative findings.
  • Exploratory sequential: Qualitative data is collected and analyzed first, followed by quantitative data. You can use this design if you think the quantitative data will confirm or validate your qualitative findings.

Triangulation in research means using multiple datasets, methods, theories and/or investigators to address a research question. It’s a research strategy that can help you enhance the validity and credibility of your findings.

Triangulation is mainly used in qualitative research , but it’s also commonly applied in quantitative research . Mixed methods research always uses triangulation.

In multistage sampling , or multistage cluster sampling, you draw a sample from a population using smaller and smaller groups at each stage.

This method is often used to collect data from a large, geographically spread group of people in national surveys, for example. You take advantage of hierarchical groupings (e.g., from state to city to neighborhood) to create a sample that’s less expensive and time-consuming to collect data from.

No, the steepness or slope of the line isn’t related to the correlation coefficient value. The correlation coefficient only tells you how closely your data fit on a line, so two datasets with the same correlation coefficient can have very different slopes.

To find the slope of the line, you’ll need to perform a regression analysis .

Correlation coefficients always range between -1 and 1.

The sign of the coefficient tells you the direction of the relationship: a positive value means the variables change together in the same direction, while a negative value means they change together in opposite directions.

The absolute value of a number is equal to the number without its sign. The absolute value of a correlation coefficient tells you the magnitude of the correlation: the greater the absolute value, the stronger the correlation.

These are the assumptions your data must meet if you want to use Pearson’s r :

  • Both variables are on an interval or ratio level of measurement
  • Data from both variables follow normal distributions
  • Your data have no outliers
  • Your data is from a random or representative sample
  • You expect a linear relationship between the two variables

Quantitative research designs can be divided into two main categories:

  • Correlational and descriptive designs are used to investigate characteristics, averages, trends, and associations between variables.
  • Experimental and quasi-experimental designs are used to test causal relationships .

Qualitative research designs tend to be more flexible. Common types of qualitative design include case study , ethnography , and grounded theory designs.

A well-planned research design helps ensure that your methods match your research aims, that you collect high-quality data, and that you use the right kind of analysis to answer your questions, utilizing credible sources . This allows you to draw valid , trustworthy conclusions.

The priorities of a research design can vary depending on the field, but you usually have to specify:

  • Your research questions and/or hypotheses
  • Your overall approach (e.g., qualitative or quantitative )
  • The type of design you’re using (e.g., a survey , experiment , or case study )
  • Your sampling methods or criteria for selecting subjects
  • Your data collection methods (e.g., questionnaires , observations)
  • Your data collection procedures (e.g., operationalization , timing and data management)
  • Your data analysis methods (e.g., statistical tests  or thematic analysis )

A research design is a strategy for answering your   research question . It defines your overall approach and determines how you will collect and analyze data.

Questionnaires can be self-administered or researcher-administered.

Self-administered questionnaires can be delivered online or in paper-and-pen formats, in person or through mail. All questions are standardized so that all respondents receive the same questions with identical wording.

Researcher-administered questionnaires are interviews that take place by phone, in-person, or online between researchers and respondents. You can gain deeper insights by clarifying questions for respondents or asking follow-up questions.

You can organize the questions logically, with a clear progression from simple to complex, or randomly between respondents. A logical flow helps respondents process the questionnaire easier and quicker, but it may lead to bias. Randomization can minimize the bias from order effects.

Closed-ended, or restricted-choice, questions offer respondents a fixed set of choices to select from. These questions are easier to answer quickly.

Open-ended or long-form questions allow respondents to answer in their own words. Because there are no restrictions on their choices, respondents can answer in ways that researchers may not have otherwise considered.

A questionnaire is a data collection tool or instrument, while a survey is an overarching research method that involves collecting and analyzing data from people using questionnaires.

The third variable and directionality problems are two main reasons why correlation isn’t causation .

The third variable problem means that a confounding variable affects both variables to make them seem causally related when they are not.

The directionality problem is when two variables correlate and might actually have a causal relationship, but it’s impossible to conclude which variable causes changes in the other.

Correlation describes an association between variables : when one variable changes, so does the other. A correlation is a statistical indicator of the relationship between variables.

Causation means that changes in one variable brings about changes in the other (i.e., there is a cause-and-effect relationship between variables). The two variables are correlated with each other, and there’s also a causal link between them.

While causation and correlation can exist simultaneously, correlation does not imply causation. In other words, correlation is simply a relationship where A relates to B—but A doesn’t necessarily cause B to happen (or vice versa). Mistaking correlation for causation is a common error and can lead to false cause fallacy .

Controlled experiments establish causality, whereas correlational studies only show associations between variables.

  • In an experimental design , you manipulate an independent variable and measure its effect on a dependent variable. Other variables are controlled so they can’t impact the results.
  • In a correlational design , you measure variables without manipulating any of them. You can test whether your variables change together, but you can’t be sure that one variable caused a change in another.

In general, correlational research is high in external validity while experimental research is high in internal validity .

A correlation is usually tested for two variables at a time, but you can test correlations between three or more variables.

A correlation coefficient is a single number that describes the strength and direction of the relationship between your variables.

Different types of correlation coefficients might be appropriate for your data based on their levels of measurement and distributions . The Pearson product-moment correlation coefficient (Pearson’s r ) is commonly used to assess a linear relationship between two quantitative variables.

A correlational research design investigates relationships between two variables (or more) without the researcher controlling or manipulating any of them. It’s a non-experimental type of quantitative research .

A correlation reflects the strength and/or direction of the association between two or more variables.

  • A positive correlation means that both variables change in the same direction.
  • A negative correlation means that the variables change in opposite directions.
  • A zero correlation means there’s no relationship between the variables.

Random error  is almost always present in scientific studies, even in highly controlled settings. While you can’t eradicate it completely, you can reduce random error by taking repeated measurements, using a large sample, and controlling extraneous variables .

You can avoid systematic error through careful design of your sampling , data collection , and analysis procedures. For example, use triangulation to measure your variables using multiple methods; regularly calibrate instruments or procedures; use random sampling and random assignment ; and apply masking (blinding) where possible.

Systematic error is generally a bigger problem in research.

With random error, multiple measurements will tend to cluster around the true value. When you’re collecting data from a large sample , the errors in different directions will cancel each other out.

Systematic errors are much more problematic because they can skew your data away from the true value. This can lead you to false conclusions ( Type I and II errors ) about the relationship between the variables you’re studying.

Random and systematic error are two types of measurement error.

Random error is a chance difference between the observed and true values of something (e.g., a researcher misreading a weighing scale records an incorrect measurement).

Systematic error is a consistent or proportional difference between the observed and true values of something (e.g., a miscalibrated scale consistently records weights as higher than they actually are).

On graphs, the explanatory variable is conventionally placed on the x-axis, while the response variable is placed on the y-axis.

  • If you have quantitative variables , use a scatterplot or a line graph.
  • If your response variable is categorical, use a scatterplot or a line graph.
  • If your explanatory variable is categorical, use a bar graph.

The term “ explanatory variable ” is sometimes preferred over “ independent variable ” because, in real world contexts, independent variables are often influenced by other variables. This means they aren’t totally independent.

Multiple independent variables may also be correlated with each other, so “explanatory variables” is a more appropriate term.

The difference between explanatory and response variables is simple:

  • An explanatory variable is the expected cause, and it explains the results.
  • A response variable is the expected effect, and it responds to other variables.

In a controlled experiment , all extraneous variables are held constant so that they can’t influence the results. Controlled experiments require:

  • A control group that receives a standard treatment, a fake treatment, or no treatment.
  • Random assignment of participants to ensure the groups are equivalent.

Depending on your study topic, there are various other methods of controlling variables .

There are 4 main types of extraneous variables :

  • Demand characteristics : environmental cues that encourage participants to conform to researchers’ expectations.
  • Experimenter effects : unintentional actions by researchers that influence study outcomes.
  • Situational variables : environmental variables that alter participants’ behaviors.
  • Participant variables : any characteristic or aspect of a participant’s background that could affect study results.

An extraneous variable is any variable that you’re not investigating that can potentially affect the dependent variable of your research study.

A confounding variable is a type of extraneous variable that not only affects the dependent variable, but is also related to the independent variable.

In a factorial design, multiple independent variables are tested.

If you test two variables, each level of one independent variable is combined with each level of the other independent variable to create different conditions.

Within-subjects designs have many potential threats to internal validity , but they are also very statistically powerful .

Advantages:

  • Only requires small samples
  • Statistically powerful
  • Removes the effects of individual differences on the outcomes

Disadvantages:

  • Internal validity threats reduce the likelihood of establishing a direct relationship between variables
  • Time-related effects, such as growth, can influence the outcomes
  • Carryover effects mean that the specific order of different treatments affect the outcomes

While a between-subjects design has fewer threats to internal validity , it also requires more participants for high statistical power than a within-subjects design .

  • Prevents carryover effects of learning and fatigue.
  • Shorter study duration.
  • Needs larger samples for high power.
  • Uses more resources to recruit participants, administer sessions, cover costs, etc.
  • Individual differences may be an alternative explanation for results.

Yes. Between-subjects and within-subjects designs can be combined in a single study when you have two or more independent variables (a factorial design). In a mixed factorial design, one variable is altered between subjects and another is altered within subjects.

In a between-subjects design , every participant experiences only one condition, and researchers assess group differences between participants in various conditions.

In a within-subjects design , each participant experiences all conditions, and researchers test the same participants repeatedly for differences between conditions.

The word “between” means that you’re comparing different conditions between groups, while the word “within” means you’re comparing different conditions within the same group.

Random assignment is used in experiments with a between-groups or independent measures design. In this research design, there’s usually a control group and one or more experimental groups. Random assignment helps ensure that the groups are comparable.

In general, you should always use random assignment in this type of experimental design when it is ethically possible and makes sense for your study topic.

To implement random assignment , assign a unique number to every member of your study’s sample .

Then, you can use a random number generator or a lottery method to randomly assign each number to a control or experimental group. You can also do so manually, by flipping a coin or rolling a dice to randomly assign participants to groups.

In experimental research, random assignment is a way of placing participants from your sample into different groups using randomization. With this method, every member of the sample has a known or equal chance of being placed in a control group or an experimental group.

“Controlling for a variable” means measuring extraneous variables and accounting for them statistically to remove their effects on other variables.

Researchers often model control variable data along with independent and dependent variable data in regression analyses and ANCOVAs . That way, you can isolate the control variable’s effects from the relationship between the variables of interest.

Control variables help you establish a correlational or causal relationship between variables by enhancing internal validity .

If you don’t control relevant extraneous variables , they may influence the outcomes of your study, and you may not be able to demonstrate that your results are really an effect of your independent variable .

A control variable is any variable that’s held constant in a research study. It’s not a variable of interest in the study, but it’s controlled because it could influence the outcomes.

Including mediators and moderators in your research helps you go beyond studying a simple relationship between two variables for a fuller picture of the real world. They are important to consider when studying complex correlational or causal relationships.

Mediators are part of the causal pathway of an effect, and they tell you how or why an effect takes place. Moderators usually help you judge the external validity of your study by identifying the limitations of when the relationship between variables holds.

If something is a mediating variable :

  • It’s caused by the independent variable .
  • It influences the dependent variable
  • When it’s taken into account, the statistical correlation between the independent and dependent variables is higher than when it isn’t considered.

A confounder is a third variable that affects variables of interest and makes them seem related when they are not. In contrast, a mediator is the mechanism of a relationship between two variables: it explains the process by which they are related.

A mediator variable explains the process through which two variables are related, while a moderator variable affects the strength and direction of that relationship.

There are three key steps in systematic sampling :

  • Define and list your population , ensuring that it is not ordered in a cyclical or periodic order.
  • Decide on your sample size and calculate your interval, k , by dividing your population by your target sample size.
  • Choose every k th member of the population as your sample.

Systematic sampling is a probability sampling method where researchers select members of the population at a regular interval – for example, by selecting every 15th person on a list of the population. If the population is in a random order, this can imitate the benefits of simple random sampling .

Yes, you can create a stratified sample using multiple characteristics, but you must ensure that every participant in your study belongs to one and only one subgroup. In this case, you multiply the numbers of subgroups for each characteristic to get the total number of groups.

For example, if you were stratifying by location with three subgroups (urban, rural, or suburban) and marital status with five subgroups (single, divorced, widowed, married, or partnered), you would have 3 x 5 = 15 subgroups.

You should use stratified sampling when your sample can be divided into mutually exclusive and exhaustive subgroups that you believe will take on different mean values for the variable that you’re studying.

Using stratified sampling will allow you to obtain more precise (with lower variance ) statistical estimates of whatever you are trying to measure.

For example, say you want to investigate how income differs based on educational attainment, but you know that this relationship can vary based on race. Using stratified sampling, you can ensure you obtain a large enough sample from each racial group, allowing you to draw more precise conclusions.

In stratified sampling , researchers divide subjects into subgroups called strata based on characteristics that they share (e.g., race, gender, educational attainment).

Once divided, each subgroup is randomly sampled using another probability sampling method.

Cluster sampling is more time- and cost-efficient than other probability sampling methods , particularly when it comes to large samples spread across a wide geographical area.

However, it provides less statistical certainty than other methods, such as simple random sampling , because it is difficult to ensure that your clusters properly represent the population as a whole.

There are three types of cluster sampling : single-stage, double-stage and multi-stage clustering. In all three types, you first divide the population into clusters, then randomly select clusters for use in your sample.

  • In single-stage sampling , you collect data from every unit within the selected clusters.
  • In double-stage sampling , you select a random sample of units from within the clusters.
  • In multi-stage sampling , you repeat the procedure of randomly sampling elements from within the clusters until you have reached a manageable sample.

Cluster sampling is a probability sampling method in which you divide a population into clusters, such as districts or schools, and then randomly select some of these clusters as your sample.

The clusters should ideally each be mini-representations of the population as a whole.

If properly implemented, simple random sampling is usually the best sampling method for ensuring both internal and external validity . However, it can sometimes be impractical and expensive to implement, depending on the size of the population to be studied,

If you have a list of every member of the population and the ability to reach whichever members are selected, you can use simple random sampling.

The American Community Survey  is an example of simple random sampling . In order to collect detailed data on the population of the US, the Census Bureau officials randomly select 3.5 million households per year and use a variety of methods to convince them to fill out the survey.

Simple random sampling is a type of probability sampling in which the researcher randomly selects a subset of participants from a population . Each member of the population has an equal chance of being selected. Data is then collected from as large a percentage as possible of this random subset.

Quasi-experimental design is most useful in situations where it would be unethical or impractical to run a true experiment .

Quasi-experiments have lower internal validity than true experiments, but they often have higher external validity  as they can use real-world interventions instead of artificial laboratory settings.

A quasi-experiment is a type of research design that attempts to establish a cause-and-effect relationship. The main difference with a true experiment is that the groups are not randomly assigned.

Blinding is important to reduce research bias (e.g., observer bias , demand characteristics ) and ensure a study’s internal validity .

If participants know whether they are in a control or treatment group , they may adjust their behavior in ways that affect the outcome that researchers are trying to measure. If the people administering the treatment are aware of group assignment, they may treat participants differently and thus directly or indirectly influence the final results.

  • In a single-blind study , only the participants are blinded.
  • In a double-blind study , both participants and experimenters are blinded.
  • In a triple-blind study , the assignment is hidden not only from participants and experimenters, but also from the researchers analyzing the data.

Blinding means hiding who is assigned to the treatment group and who is assigned to the control group in an experiment .

A true experiment (a.k.a. a controlled experiment) always includes at least one control group that doesn’t receive the experimental treatment.

However, some experiments use a within-subjects design to test treatments without a control group. In these designs, you usually compare one group’s outcomes before and after a treatment (instead of comparing outcomes between different groups).

For strong internal validity , it’s usually best to include a control group if possible. Without a control group, it’s harder to be certain that the outcome was caused by the experimental treatment and not by other variables.

An experimental group, also known as a treatment group, receives the treatment whose effect researchers wish to study, whereas a control group does not. They should be identical in all other ways.

Individual Likert-type questions are generally considered ordinal data , because the items have clear rank order, but don’t have an even distribution.

Overall Likert scale scores are sometimes treated as interval data. These scores are considered to have directionality and even spacing between them.

The type of data determines what statistical tests you should use to analyze your data.

A Likert scale is a rating scale that quantitatively assesses opinions, attitudes, or behaviors. It is made up of 4 or more questions that measure a single attitude or trait when response scores are combined.

To use a Likert scale in a survey , you present participants with Likert-type questions or statements, and a continuum of items, usually with 5 or 7 possible responses, to capture their degree of agreement.

In scientific research, concepts are the abstract ideas or phenomena that are being studied (e.g., educational achievement). Variables are properties or characteristics of the concept (e.g., performance at school), while indicators are ways of measuring or quantifying variables (e.g., yearly grade reports).

The process of turning abstract concepts into measurable variables and indicators is called operationalization .

There are various approaches to qualitative data analysis , but they all share five steps in common:

  • Prepare and organize your data.
  • Review and explore your data.
  • Develop a data coding system.
  • Assign codes to the data.
  • Identify recurring themes.

The specifics of each step depend on the focus of the analysis. Some common approaches include textual analysis , thematic analysis , and discourse analysis .

There are five common approaches to qualitative research :

  • Grounded theory involves collecting data in order to develop new theories.
  • Ethnography involves immersing yourself in a group or organization to understand its culture.
  • Narrative research involves interpreting stories to understand how people make sense of their experiences and perceptions.
  • Phenomenological research involves investigating phenomena through people’s lived experiences.
  • Action research links theory and practice in several cycles to drive innovative changes.

Hypothesis testing is a formal procedure for investigating our ideas about the world using statistics. It is used by scientists to test specific predictions, called hypotheses , by calculating how likely it is that a pattern or relationship between variables could have arisen by chance.

Operationalization means turning abstract conceptual ideas into measurable observations.

For example, the concept of social anxiety isn’t directly observable, but it can be operationally defined in terms of self-rating scores, behavioral avoidance of crowded places, or physical anxiety symptoms in social situations.

Before collecting data , it’s important to consider how you will operationalize the variables that you want to measure.

When conducting research, collecting original data has significant advantages:

  • You can tailor data collection to your specific research aims (e.g. understanding the needs of your consumers or user testing your website)
  • You can control and standardize the process for high reliability and validity (e.g. choosing appropriate measurements and sampling methods )

However, there are also some drawbacks: data collection can be time-consuming, labor-intensive and expensive. In some cases, it’s more efficient to use secondary data that has already been collected by someone else, but the data might be less reliable.

Data collection is the systematic process by which observations or measurements are gathered in research. It is used in many different contexts by academics, governments, businesses, and other organizations.

There are several methods you can use to decrease the impact of confounding variables on your research: restriction, matching, statistical control and randomization.

In restriction , you restrict your sample by only including certain subjects that have the same values of potential confounding variables.

In matching , you match each of the subjects in your treatment group with a counterpart in the comparison group. The matched subjects have the same values on any potential confounding variables, and only differ in the independent variable .

In statistical control , you include potential confounders as variables in your regression .

In randomization , you randomly assign the treatment (or independent variable) in your study to a sufficiently large number of subjects, which allows you to control for all potential confounding variables.

A confounding variable is closely related to both the independent and dependent variables in a study. An independent variable represents the supposed cause , while the dependent variable is the supposed effect . A confounding variable is a third variable that influences both the independent and dependent variables.

Failing to account for confounding variables can cause you to wrongly estimate the relationship between your independent and dependent variables.

To ensure the internal validity of your research, you must consider the impact of confounding variables. If you fail to account for them, you might over- or underestimate the causal relationship between your independent and dependent variables , or even find a causal relationship where none exists.

Yes, but including more than one of either type requires multiple research questions .

For example, if you are interested in the effect of a diet on health, you can use multiple measures of health: blood sugar, blood pressure, weight, pulse, and many more. Each of these is its own dependent variable with its own research question.

You could also choose to look at the effect of exercise levels as well as diet, or even the additional effect of the two combined. Each of these is a separate independent variable .

To ensure the internal validity of an experiment , you should only change one independent variable at a time.

No. The value of a dependent variable depends on an independent variable, so a variable cannot be both independent and dependent at the same time. It must be either the cause or the effect, not both!

You want to find out how blood sugar levels are affected by drinking diet soda and regular soda, so you conduct an experiment .

  • The type of soda – diet or regular – is the independent variable .
  • The level of blood sugar that you measure is the dependent variable – it changes depending on the type of soda.

Determining cause and effect is one of the most important parts of scientific research. It’s essential to know which is the cause – the independent variable – and which is the effect – the dependent variable.

In non-probability sampling , the sample is selected based on non-random criteria, and not every member of the population has a chance of being included.

Common non-probability sampling methods include convenience sampling , voluntary response sampling, purposive sampling , snowball sampling, and quota sampling .

Probability sampling means that every member of the target population has a known chance of being included in the sample.

Probability sampling methods include simple random sampling , systematic sampling , stratified sampling , and cluster sampling .

Using careful research design and sampling procedures can help you avoid sampling bias . Oversampling can be used to correct undercoverage bias .

Some common types of sampling bias include self-selection bias , nonresponse bias , undercoverage bias , survivorship bias , pre-screening or advertising bias, and healthy user bias.

Sampling bias is a threat to external validity – it limits the generalizability of your findings to a broader group of people.

A sampling error is the difference between a population parameter and a sample statistic .

A statistic refers to measures about the sample , while a parameter refers to measures about the population .

Populations are used when a research question requires data from every member of the population. This is usually only feasible when the population is small and easily accessible.

Samples are used to make inferences about populations . Samples are easier to collect data from because they are practical, cost-effective, convenient, and manageable.

There are seven threats to external validity : selection bias , history, experimenter effect, Hawthorne effect , testing effect, aptitude-treatment and situation effect.

The two types of external validity are population validity (whether you can generalize to other groups of people) and ecological validity (whether you can generalize to other situations and settings).

The external validity of a study is the extent to which you can generalize your findings to different groups of people, situations, and measures.

Cross-sectional studies cannot establish a cause-and-effect relationship or analyze behavior over a period of time. To investigate cause and effect, you need to do a longitudinal study or an experimental study .

Cross-sectional studies are less expensive and time-consuming than many other types of study. They can provide useful insights into a population’s characteristics and identify correlations for further research.

Sometimes only cross-sectional data is available for analysis; other times your research question may only require a cross-sectional study to answer it.

Longitudinal studies can last anywhere from weeks to decades, although they tend to be at least a year long.

The 1970 British Cohort Study , which has collected data on the lives of 17,000 Brits since their births in 1970, is one well-known example of a longitudinal study .

Longitudinal studies are better to establish the correct sequence of events, identify changes over time, and provide insight into cause-and-effect relationships, but they also tend to be more expensive and time-consuming than other types of studies.

Longitudinal studies and cross-sectional studies are two different types of research design . In a cross-sectional study you collect data from a population at a specific point in time; in a longitudinal study you repeatedly collect data from the same sample over an extended period of time.

Longitudinal study Cross-sectional study
observations Observations at a in time
Observes the multiple times Observes (a “cross-section”) in the population
Follows in participants over time Provides of society at a given point

There are eight threats to internal validity : history, maturation, instrumentation, testing, selection bias , regression to the mean, social interaction and attrition .

Internal validity is the extent to which you can be confident that a cause-and-effect relationship established in a study cannot be explained by other factors.

In mixed methods research , you use both qualitative and quantitative data collection and analysis methods to answer your research question .

The research methods you use depend on the type of data you need to answer your research question .

  • If you want to measure something or test a hypothesis , use quantitative methods . If you want to explore ideas, thoughts and meanings, use qualitative methods .
  • If you want to analyze a large amount of readily-available data, use secondary data. If you want data specific to your purposes with control over how it is generated, collect primary data.
  • If you want to establish cause-and-effect relationships between variables , use experimental methods. If you want to understand the characteristics of a research subject, use descriptive methods.

A confounding variable , also called a confounder or confounding factor, is a third variable in a study examining a potential cause-and-effect relationship.

A confounding variable is related to both the supposed cause and the supposed effect of the study. It can be difficult to separate the true effect of the independent variable from the effect of the confounding variable.

In your research design , it’s important to identify potential confounding variables and plan how you will reduce their impact.

Discrete and continuous variables are two types of quantitative variables :

  • Discrete variables represent counts (e.g. the number of objects in a collection).
  • Continuous variables represent measurable amounts (e.g. water volume or weight).

Quantitative variables are any variables where the data represent amounts (e.g. height, weight, or age).

Categorical variables are any variables where the data represent groups. This includes rankings (e.g. finishing places in a race), classifications (e.g. brands of cereal), and binary outcomes (e.g. coin flips).

You need to know what type of variables you are working with to choose the right statistical test for your data and interpret your results .

You can think of independent and dependent variables in terms of cause and effect: an independent variable is the variable you think is the cause , while a dependent variable is the effect .

In an experiment, you manipulate the independent variable and measure the outcome in the dependent variable. For example, in an experiment about the effect of nutrients on crop growth:

  • The  independent variable  is the amount of nutrients added to the crop field.
  • The  dependent variable is the biomass of the crops at harvest time.

Defining your variables, and deciding how you will manipulate and measure them, is an important part of experimental design .

Experimental design means planning a set of procedures to investigate a relationship between variables . To design a controlled experiment, you need:

  • A testable hypothesis
  • At least one independent variable that can be precisely manipulated
  • At least one dependent variable that can be precisely measured

When designing the experiment, you decide:

  • How you will manipulate the variable(s)
  • How you will control for any potential confounding variables
  • How many subjects or samples will be included in the study
  • How subjects will be assigned to treatment levels

Experimental design is essential to the internal and external validity of your experiment.

I nternal validity is the degree of confidence that the causal relationship you are testing is not influenced by other factors or variables .

External validity is the extent to which your results can be generalized to other contexts.

The validity of your experiment depends on your experimental design .

Reliability and validity are both about how well a method measures something:

  • Reliability refers to the  consistency of a measure (whether the results can be reproduced under the same conditions).
  • Validity   refers to the  accuracy of a measure (whether the results really do represent what they are supposed to measure).

If you are doing experimental research, you also have to consider the internal and external validity of your experiment.

A sample is a subset of individuals from a larger population . Sampling means selecting the group that you will actually collect data from in your research. For example, if you are researching the opinions of students in your university, you could survey a sample of 100 students.

In statistics, sampling allows you to test a hypothesis about the characteristics of a population.

Quantitative research deals with numbers and statistics, while qualitative research deals with words and meanings.

Quantitative methods allow you to systematically measure variables and test hypotheses . Qualitative methods allow you to explore concepts and experiences in more detail.

Methodology refers to the overarching strategy and rationale of your research project . It involves studying the methods used in your field and the theories or principles behind them, in order to develop an approach that matches your objectives.

Methods are the specific tools and procedures you use to collect and analyze data (for example, experiments, surveys , and statistical tests ).

In shorter scientific papers, where the aim is to report the findings of a specific study, you might simply describe what you did in a methods section .

In a longer or more complex research project, such as a thesis or dissertation , you will probably include a methodology section , where you explain your approach to answering the research questions and cite relevant sources to support your choice of methods.

Ask our team

Want to contact us directly? No problem.  We  are always here for you.

Support team - Nina

Our team helps students graduate by offering:

  • A world-class citation generator
  • Plagiarism Checker software powered by Turnitin
  • Innovative Citation Checker software
  • Professional proofreading services
  • Over 300 helpful articles about academic writing, citing sources, plagiarism, and more

Scribbr specializes in editing study-related documents . We proofread:

  • PhD dissertations
  • Research proposals
  • Personal statements
  • Admission essays
  • Motivation letters
  • Reflection papers
  • Journal articles
  • Capstone projects

Scribbr’s Plagiarism Checker is powered by elements of Turnitin’s Similarity Checker , namely the plagiarism detection software and the Internet Archive and Premium Scholarly Publications content databases .

The add-on AI detector is powered by Scribbr’s proprietary software.

The Scribbr Citation Generator is developed using the open-source Citation Style Language (CSL) project and Frank Bennett’s citeproc-js . It’s the same technology used by dozens of other popular citation tools, including Mendeley and Zotero.

You can find all the citation styles and locales used in the Scribbr Citation Generator in our publicly accessible repository on Github .

Department of Health & Human Services

Module 3: Elements of Research - Section 4

Module 1

Section 4: Random Selection

Groups of people

Definition : Random selection is the process of selecting a smaller group of individuals from a larger group to be participants in a study. Every person has an equal chance of being selected, which allows each of the individuals in the group the same chance of participating.

Case Example for Random Selection

Names are drawn in a hat. Special computer programs

Section 4: Discussion Questions

  • Do the women included on this list represent the larger group?
  • Would you choose participants differently if this was your study? If so, how would you do it?
  • What if the researcher decided to ignore the random selection procedures as described in this example (e.g. pulling names from a hat, using a computer program to generate random numbers)? How would this affect the results of the study?

PDF

Email Updates

Information

  • Author Services

Initiatives

You are accessing a machine-readable page. In order to be human-readable, please install an RSS reader.

All articles published by MDPI are made immediately available worldwide under an open access license. No special permission is required to reuse all or part of the article published by MDPI, including figures and tables. For articles published under an open access Creative Common CC BY license, any part of the article may be reused without permission provided that the original article is clearly cited. For more information, please refer to https://www.mdpi.com/openaccess .

Feature papers represent the most advanced research with significant potential for high impact in the field. A Feature Paper should be a substantial original Article that involves several techniques or approaches, provides an outlook for future research directions and describes possible research applications.

Feature papers are submitted upon individual invitation or recommendation by the scientific editors and must receive positive feedback from the reviewers.

Editor’s Choice articles are based on recommendations by the scientific editors of MDPI journals from around the world. Editors select a small number of articles recently published in the journal that they believe will be particularly interesting to readers, or important in the respective research area. The aim is to provide a snapshot of some of the most exciting work published in the various research areas of the journal.

Original Submission Date Received: .

  • Active Journals
  • Find a Journal
  • Proceedings Series
  • For Authors
  • For Reviewers
  • For Editors
  • For Librarians
  • For Publishers
  • For Societies
  • For Conference Organizers
  • Open Access Policy
  • Institutional Open Access Program
  • Special Issues Guidelines
  • Editorial Process
  • Research and Publication Ethics
  • Article Processing Charges
  • Testimonials
  • Preprints.org
  • SciProfiles
  • Encyclopedia

animals-logo

Article Menu

randomly selected in research

  • Subscribe SciFeed
  • Recommended Articles
  • Author Biographies
  • Google Scholar
  • on Google Scholar
  • Table of Contents

Find support for a specific problem in the support section of our website.

Please let us know what you think of our products and services.

Visit our dedicated information section to learn more about MDPI.

JSmol Viewer

Investigating the reasons behind companion animal relinquishment: a systematic content analysis of shelter records for cats and dogs, 2018–2023.

randomly selected in research

Simple Summary

1. introduction, 2. materials and method, 2.1. shelter, species, and time frame, 2.2. relinquishment records, 2.3. content analysis, 2.4. data analysis, 3.1. comparison of species: cats vs. dogs, 3.2. comparison of type: first relinquishment vs. return, 3.3. comparison of years: 2018–2023, 4. discussion, 4.1. finding highlights, 4.2. limitations and future directions, 5. conclusions, author contributions, institutional review board statement, informed consent statement, data availability statement, acknowledgments, conflicts of interest.

  • Blazina, C.; Boyraz, G.; Shen-Miller, D. (Eds.) The Psychology of the Human-Animal Bond ; Springer: New York, NY, USA, 2011; ISBN 978-1-4419-9760-9. [ Google Scholar ]
  • Hines, L.M. Historical Perspectives on the Human-Animal Bond. Am. Behav. Sci. 2003 , 47 , 7–15. [ Google Scholar ] [ CrossRef ]
  • Hosey, G.; Melfi, V. Human-Animal Interactions, Relationships and Bonds: A Review and Analysis of the Literature. IJCP 2014 , 27 , 117–142. [ Google Scholar ] [ CrossRef ]
  • Irvine, L. If You Tame Me: Understanding Our Connection with Animals ; Temple University Press: Philadelphia, PA, USA, 2004. [ Google Scholar ]
  • Serpell, J. The Company of Animals: A Study of Human-Animal Relationships ; Cambridge University Press: Cambridge, MA, USA, 1996; ISBN 978-0-521-57779-3. [ Google Scholar ]
  • Amiot, C.E.; Bastian, B. Toward a Psychology of Human–Animal Relations. Psychol. Bull. 2015 , 141 , 6–47. [ Google Scholar ] [ CrossRef ]
  • McCune, S.; McCardle, P.; Griffin, J.A.; Esposito, L.; Hurley, K.; Bures, R.; Kruger, K.A. Editorial: Human-Animal Interaction (HAI) Research: A Decade of Progress. Front. Vet. Sci. 2020 , 7 , 44. [ Google Scholar ] [ CrossRef ]
  • Rault, J.-L.; Waiblinger, S.; Boivin, X.; Hemsworth, P. The Power of a Positive Human–Animal Relationship for Animal Welfare. Front. Vet. Sci. 2020 , 7 , 590867. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Wells, D.L. The Effects of Animals on Human Health and Well-Being. J. Soc. Issues 2009 , 65 , 523–543. [ Google Scholar ] [ CrossRef ]
  • Coe, J.B.; Young, I.; Lambert, K.; Dysart, L.; Nogueira Borden, L.; Rajić, A. A Scoping Review of Published Research on the Relinquishment of Companion Animals. J. Appl. Anim. Welf. Sci. 2014 , 17 , 253–273. [ Google Scholar ] [ CrossRef ]
  • Lambert, K.; Coe, J.; Niel, L.; Dewey, C.; Sargeant, J.M. A Systematic Review and Meta-Analysis of the Proportion of Dogs Surrendered for Dog-Related and Owner-Related Reasons. Prev. Vet. Med. 2015 , 118 , 148–160. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Shelter Animals Count National Animal Welfare Statistics. Available online: https://www.shelteranimalscount.org/explore-the-data/data-dashboards/national-animal-welfare-statistics-dashboard/ (accessed on 1 July 2024).
  • Bradley, J.; Patronek, G.J. Dog Behavior and Relinquishment to Shelters. In Animal Behavior for Shelter Veterinarians and Staff ; DiGangi, B.A., Cussen, V.A., Reid, P.J., Collins, K.A., Eds.; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 2022; pp. 133–152. ISBN 978-1-119-61851-5. [ Google Scholar ]
  • Thumpkin, E.; Paterson, M.B.A.; Morton, J.M.; Pachana, N.A. Adoption Can Be a Risky Business: Risk Factors Predictive of Dogs Adopted from RSPCA Queensland Being Returned. Animals 2022 , 12 , 2568. [ Google Scholar ] [ CrossRef ]
  • Weiss, E.; Slater, M.; Garrison, L.; Drain, N.; Dolan, E.; Scarlett, J.; Zawistowski, S. Large Dog Relinquishment to Two Municipal Facilities in New York City and Washington, D.C.: Identifying Targets for Intervention. Animals 2014 , 4 , 409–433. [ Google Scholar ] [ CrossRef ]
  • Sharkin, B.S.; Ruff, L.A. Broken Bonds: Understanding the Experience of Pet Relinquishment. In The Psychology of the Human-Animal Bond A Resource for Clinicians and Researchers ; Blazina, C., Boyraz, G., Shen-Miller, D., Eds.; Springer: New York, NY, USA, 2011; pp. 275–287. [ Google Scholar ]
  • Mundschau, V.; Suchak, M. When and Why Cats Are Returned to Shelters. Animals 2023 , 13 , 243. [ Google Scholar ] [ CrossRef ]
  • Miller, D.D.; Staats, S.R.; Partlo, C.; Rada, K. Factors Associated with the Decision to Surrender a Pet to an Animal Shelter. Javma 1996 , 209 , 738–742. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Diesel, G.; Brodbelt, D.; Pfeiffer, D.U. Characteristics of Relinquished Dogs and Their Owners at 14 Rehoming Centers in the United Kingdom. J. Appl. Anim. Welf. Sci. 2010 , 13 , 15–30. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Miller, H.; Ward, M.; Beatty, J.A. Population Characteristics of Cats Adopted from an Urban Cat Shelter and the Influence of Physical Traits and Reason for Surrender on Length of Stay. Animals 2019 , 9 , 940. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Alberthsen, C.; Rand, J.; Morton, J.; Bennett, P.; Paterson, M.; Vankan, D. Numbers and Characteristics of Cats Admitted to Royal Society for the Prevention of Cruelty to Animals (RSPCA) Shelters in Australia and Reasons for Surrender. Animals 2016 , 6 , 23. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Scarlett, J.M.; Salman, M.D.; New, J.G.; Kass, P.H. Reasons for Relinquishment of Companion Animals in U.S. Animal Shelters: Selected Health and Personal Issues. J. Appl. Anim. Welf. Sci. 1999 , 2 , 41–57. [ Google Scholar ] [ CrossRef ]
  • Dolan, E.; Scotto, J.; Slater, M.; Weiss, E. Risk Factors for Dog Relinquishment to a Los Angeles Municipal Animal Shelter. Animals 2015 , 5 , 1311–1328. [ Google Scholar ] [ CrossRef ]
  • Jensen, J.B.H.; Sandøe, P.; Nielsen, S.S. Owner-Related Reasons Matter More than Behavioural Problems—A Study of Why Owners Relinquished Dogs and Cats to a Danish Animal Shelter from 1996 to 2017. Animals 2020 , 10 , 1064. [ Google Scholar ] [ CrossRef ]
  • Rah, H.; Choi, S.-H. Are Veterinary Costs and Socioeconomic Status Risk Factors for Companion Animal Relinquishment in the Republic of Korea? Animals 2023 , 13 , 3406. [ Google Scholar ] [ CrossRef ]
  • Crano, W.D.; Brewer, M.B.; Lac, A. Principles and Methods of Social Research , 3rd ed.; Routledge: New York, NY, USA, 2014; ISBN 978-1-315-76831-1. [ Google Scholar ]
  • Stemler, S. An Overview of Content Analysis. Pract. Assess. Res. Eval. 2001 , 7 , 1–6. [ Google Scholar ] [ CrossRef ]
  • Brand, C.L.; O’Neill, D.G.; Belshaw, Z.; Dale, F.C.; Merritt, B.L.; Clover, K.N.; Tay, M.-X.M.; Pegram, C.L.; Packer, R.M.A. Impacts of Puppy Early Life Experiences, Puppy-Purchasing Practices, and Owner Characteristics on Owner-Reported Problem Behaviours in a UK Pandemic Puppies Cohort at 21 Months of Age. Animals 2024 , 14 , 336. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Powell, L.; Reinhard, C.; Satriale, D.; Morris, M.; Serpell, J.; Watson, B. Characterizing Unsuccessful Animal Adoptions: Age and Breed Predict the Likelihood of Return, Reasons for Return and Post-Return Outcomes. Sci. Rep. 2021 , 11 , 8018. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Hawes, S.M.; Kerrigan, J.M.; Hupe, T.; Morris, K.N. Factors Informing the Return of Adopted Dogs and Cats to an Animal Shelter. Animals 2020 , 10 , 1573. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Morgan, L.; Protopopova, A.; Birkler, R.I.D.; Itin-Shwartz, B.; Sutton, G.A.; Gamliel, A.; Yakobson, B.; Raz, T. Human–Dog Relationships during the COVID-19 Pandemic: Booming Dog Adoption during Social Isolation. Humanit. Soc. Sci. Commun. 2020 , 7 , 155. [ Google Scholar ] [ CrossRef ]
  • Ogata, N.; Weng, H.-Y.; Messam, L.L.M. Temporal Patterns of Owner-Pet Relationship, Stress, and Loneliness during the COVID-19 Pandemic, and the Effect of Pet Ownership on Mental Health: A Longitudinal Survey. PLoS ONE 2023 , 18 , e0284101. [ Google Scholar ] [ CrossRef ]
  • Powell, L.; Houlihan, C.; Stone, M.; Gitlin, I.; Ji, X.; Reinhard, C.L.; Watson, B. Animal Shelters’ Response to the COVID-19 Pandemic: A Pilot Survey of 14 Shelters in the Northeastern United States. Animals 2021 , 11 , 2669. [ Google Scholar ] [ CrossRef ]
  • Rodriguez, J.R.; Davis, J.; Hill, S.; Wolf, P.J.; Hawes, S.M.; Morris, K.N. Trends in Intake and Outcome Data From, U.S. Animal Shelters From 2016 to 2020. Front. Vet. Sci. 2022 , 9 , 863990. [ Google Scholar ] [ CrossRef ]
  • Carroll, G.A.; Torjussen, A.; Reeve, C. Companion Animal Adoption and Relinquishment during the COVID-19 Pandemic: Peri-Pandemic Pets at Greatest Risk of Relinquishment. Front. Vet. Sci. 2022 , 9 , 1017954. [ Google Scholar ] [ CrossRef ]
  • Carroll, G.A.; Torjussen, A.; Reeve, C. Companion Animal Adoption and Relinquishment during the COVID-19 Pandemic: Households with Children at Greatest Risk of Relinquishing a Cat or Dog. Anim. Welf. 2023 , 32 , e56. [ Google Scholar ] [ CrossRef ]
  • Asimolar Accords Asimolar Accords. Saving Lives Not Pointing Fingers. 2004. Available online: https://www.americanhumane.org/app/uploads/2016/08/au-asilomar-accordspdf.pdf (accessed on 1 July 2024).
  • Cicchetti, D.V. Guidelines, Criteria, and Rules of Thumb for Evaluating Normed and Standardized Assessment Instruments in Psychology. Psychol. Assess. 1994 , 6 , 284–290. [ Google Scholar ] [ CrossRef ]
  • Cochran, W.G. The Chi-Square Test of Goodness of Fit. Ann. Math. Stat. 1952 , 23 , 315–345. [ Google Scholar ] [ CrossRef ]
  • Hallgren, K.A. Computing Inter-Rater Reliability for Observational Data: An Overview and Tutorial. TQMP 2012 , 8 , 23–34. [ Google Scholar ] [ CrossRef ]
  • Lac, A. Content Analysis. In Encyclopedia of Adolescence ; Levesque, R.J.R., Ed.; Springer International Publishing: Cham, Switzerland, 2016; pp. 1–5. ISBN 978-3-319-32132-5. [ Google Scholar ]
  • Lombard, M.; Snyder-Duch, J.; Bracken, C.C. Content Analysis in Mass Communication: Assessment and Reporting of Intercoder Reliability. Hum. Commun. Res. 2002 , 28 , 587–604. [ Google Scholar ] [ CrossRef ]
  • McHugh, M.L. The Chi-Square Test of Independence. Biochem. Med. 2013 , 23 , 143–149. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Sharpe, D. Chi-Square Test Is Statistically Significant: Now What. Pract. Assess. Res. Eval. 2015 , 20 , 1–10. [ Google Scholar ] [ CrossRef ]
  • DiGiacomo, N.; Arluke, A.; Patronek, G. Surrendering Pets To Shelters: The Relinquisher’s Perspective. Anthrozoös 1998 , 11 , 41–51. [ Google Scholar ] [ CrossRef ]
  • Kwan, J.Y.; Bain, M.J. Owner Attachment and Problem Behaviors Related to Relinquishment and Training Techniques of Dogs. J. Appl. Anim. Welf. Sci. 2013 , 16 , 168–183. [ Google Scholar ] [ CrossRef ]
  • Bouma, E.M.C.; Dijkstra, A.; Arnt Rosa, S. Owner’s Anthropomorphic Perceptions of Cats’ and Dogs’ Abilities Are Related to the Social Role of Pets, Owners’ Relationship Behaviors, and Social Support. Animals 2023 , 13 , 3644. [ Google Scholar ] [ CrossRef ]
  • Menchetti, L.; Calipari, S.; Guelfi, G.; Catanzaro, A.; Diverio, S. My Dog Is Not My Cat: Owner Perception of the Personalities of Dogs and Cats Living in the Same Household. Animals 2018 , 8 , 80. [ Google Scholar ] [ CrossRef ]
  • Fatjó, J.; Ruiz-de-la-Torre, J.; Manteca, X. The Epidemiology of Behavioural Problems in Dogs and Cats: A Survey of Veterinary Practitioners. Anim. Welf. 2006 , 15 , 179–185. [ Google Scholar ] [ CrossRef ]
  • Diesel, G.; Pfeiffer, D.U.; Brodbelt, D. Factors Affecting the Success of Rehoming Dogs in the UK during 2005. Prev. Vet. Med. 2008 , 84 , 228–241. [ Google Scholar ] [ CrossRef ]
  • Shore, E.R. Returning a Recently Adopted Companion Animal: Adopters’ Reasons for and Reactions to the Failed Adoption Experience. J. Appl. Anim. Welf. Sci. 2005 , 8 , 187–198. [ Google Scholar ] [ CrossRef ]
  • Wells, D.L.; Hepper, P.G. Prevalence of Behaviour Problems Reported by Owners of Dogs Purchased from an Animal Rescue Shelter. Appl. Anim. Behav. Sci. 2000 , 69 , 55–65. [ Google Scholar ] [ CrossRef ]
  • Protopopova, A.; Gunter, L. Adoption and Relinquishment Interventions at the Animal Shelter: A Review. Anim. Welf. 2017 , 26 , 35–48. [ Google Scholar ] [ CrossRef ]
  • Powdrill-Wells, N.; Taylor, S.; Melfi, V. Reducing Dog Relinquishment to Rescue Centres Due to Behaviour Problems: Identifying Cases to Target with an Advice Intervention at the Point of Relinquishment Request. Animals 2021 , 11 , 2766. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Eagan, B.H.; Gordon, E.; Protopopova, A. Reasons for Guardian-Relinquishment of Dogs to Shelters: Animal and Regional Predictors in British Columbia, Canada. Front. Vet. Sci. 2022 , 9 , 857634. [ Google Scholar ] [ CrossRef ]
  • Salman, M.D.; New, J.G.; Scarlett, J.M.; Kass, P.H.; Ruch-Gallie, R.; Hetts, S. Human and Animal Factors Related to Relinquishment of Dogs and Cats in 12 Selected Animal Shelters in the United States. J. Appl. Anim. Welf. Sci. 1998 , 1 , 207–226. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Hosley, R. How Pet Benefits Can Help Employees Return to the Office. Available online: https://www.care.com/business/how-pet-benefits-can-help-employees-return-to-the-office/ (accessed on 3 July 2024).
  • Kawamoto, D. The Latest Trend in Employee Retention? Even More Pet Benefits. Available online: https://hrexecutive.com/the-latest-trend-in-employee-retention-even-more-pet-benefits/ (accessed on 3 July 2024).
  • American Veterinary Medical Association, U.S. Pet Ownership Statistics. Available online: https://www.avma.org/resources-tools/reports-statistics/us-pet-ownership-statistics. (accessed on 1 July 2024).
  • New, J.C.; Kelch, W.J.; Hutchison, J.M.; Salman, M.D.; King, M.; Scarlett, J.M.; Kass, P.H. Birth and Death Rate Estimates of Cats and Dogs in U.S. Households and Related Factors. J. Appl. Anim. Welf. Sci. 2004 , 7 , 229–241. [ Google Scholar ] [ CrossRef ]
  • Chan, S.K.; Leung, D.Y.M. Dog and Cat Allergies: Current State of Diagnostic Approaches and Challenges. Allergy Asthma Immunol. Res. 2018 , 10 , 97. [ Google Scholar ] [ CrossRef ]
  • Ipsos Most Americans Still Frequently Washing Their Hands, but Hand Hygiene Is Less Top of Mind than When COVID-19 Began. Available online: https://www.ipsos.com/en-us/news-polls/most-americans-still-frequently-washing-their-hands-but-hand-hygiene-is-less-top-of-mind-than-when-covid-19-began. (accessed on 1 July 2024).
  • Koo, M.J.; Cho, S.H. Allergic Diseases and COVID-19 Pandemic: Facing Unprecedented Challenges. Life Cycle 2023 , 3 , e5. [ Google Scholar ] [ CrossRef ]
  • U.S. Department of Defense Update to Conditions-Based Approach to Coronavirus Disease 2019 Personnel Movement and Travel Restrictions. Available online: https://www.defense.gov/News/Releases/Release/Article/2538641/update-to-conditions-based-approach-to-coronavirus-disease-2019-personnel-movem/. (accessed on 1 July 2024).
  • U.S. Secretary of Defense Travel Restrictions for DoD Components in Response to Coronavirus Disease 2019. Available online: https://media.defense.gov/2020/Mar/11/2002263242/-1/-1/1/TRAVEL-RESTRICTIONS-FOR-DOD-COMPONENTS-IN-RESPONSE-TO-CORONAVIRUS-DISEASE-2019.PDF. (accessed on 1 July 2024).
  • Powell, L.; Duffy, D.L.; Kruger, K.A.; Watson, B.; Serpell, J.A. Relinquishing Owners Underestimate Their Dog’s Behavioral Problems: Deception or Lack of Knowledge? Front. Vet. Sci. 2021 , 8 , 734973. [ Google Scholar ] [ CrossRef ] [ PubMed ]

Click here to enlarge figure

DogsCatsTotalIRR
YearReportedAnalyzedReportedAnalyzedReportedAnalyzedOverall ReasonsBehavior Reasons
20182934243267525356094960.800.76
20192932223302219659544190.770.76
20202245219258724548324640.820.80
20212707222346925261764740.820.84
20223604229433326279374910.830.83
20234234252405824082924920.830.79
CodeCategoryDescription
1.FinancialAny mention of insufficient financial resources to care for an animal, including medical care, cost of food, or any other reason, as long as money was mentioned.
2.Sick AnimalThe owner cited the animal’s illness. If cost of care was mentioned, this would instead be coded as 1. Financial.
3.Owner AllergiesApplied whether it was the owner themselves or other family members experiencing allergy symptoms due to the animal.
4.No OwnerThe individual relinquishing the animal either found the animal or received it from another person or source, believing the animal was without an owner.
5.Unable to CareThe owner felt unable to care for the animal for any reason except financial. Common reasons included lack of time, being away from home too much, major life changes (not financial or housing-related), or a desire for the animal to have a better quality of life.
6.Do Not WantThe owner did not mention any challenges to caring for the animal but did not want the animal nevertheless.
7.Military-RelatedWhen the stated reason was related to military issues such as transfer or deployment.
8.Housing/MovingThe owner’s housing situation had changed in some way that impacted the animal, but not military-related. Common reasons included moving, breed restrictions, new landlord rules at the owner’s rental property, or becoming unhoused.
9.CKC ProgramThe Community Kitten Care (CKC) Program launched in 2020 at the Pueblo location, allowing community members providing foster care to previously feral kittens to relinquish the animals once they were 8 weeks and 2 pounds. This is separate from the animal foster program that spans both locations.
10.Behavior IssuesSelected if the animal’s behavior was the first reason cited. A subsequent behavioral code was then selected in a second step (see ).
11.DeathThe passing of the owner, a family member, a significant other, or someone close.
12.Too Many PetsThe stated reason included mentioned too many pets, the need to reduce the number of animals, unplanned litter, or other reasons for inability or lack of desire to care for so many animals. If money was mentioned, this was instead coded as 1. Financial.
13.OtherAny stated reason that did not fit with one of the above categories or was too vague or unclear to be coded.
CodeCategoryDescription
A.None givenThe owner did not provide any details beyond mention of the animal’s behavior.
B.AggressionAny mention of hostile behavior, including but not limited to causing harm to humans or other animals. Frequent behaviors mentioned included growling, biting, scratching, and any behaviors that were perceived as ‘scary’, ‘intimidating’, ‘reactive’, ‘too rough’, or similar.
C.Social ConflictNon-aggressive behaviors that nevertheless were perceived as arising from conflict with other animals or humans. This included showing fear of others, ‘not liking’ or ‘not getting along with’ others.
D.SoilingAnimal eliminated outside of desired areas.
E.Too NoisyNoise made by the animal was specifically mentioned. This usually included barking, but sometimes cat mewing as well.
F.Escape BehaviorsThe primary reason cited related to the animal attempting to break free from living space. Common behaviors included running through open doors and jumping fences.
G.Destructive BehaviorsThe animal was causing damage to property.
H.AnxietyThe animal showed behavior that can be classified as nervous, worried, or tense, including separation anxiety. If fearful behavior was induced by the presence of another animal, that would have been coded as 2. Social Conflict.
I.Too EnergeticThe animal showed great energy or vitality. Common citations typically included ‘too much’ energy, strength, or play behaviors.
J.OtherAny stated reason that did not fit with one of the above categories or was too vague or unclear to be coded.
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

Kisley, M.A.; Chung, E.J.; Levitt, H. Investigating the Reasons behind Companion Animal Relinquishment: A Systematic Content Analysis of Shelter Records for Cats and Dogs, 2018–2023. Animals 2024 , 14 , 2606. https://doi.org/10.3390/ani14172606

Kisley MA, Chung EJ, Levitt H. Investigating the Reasons behind Companion Animal Relinquishment: A Systematic Content Analysis of Shelter Records for Cats and Dogs, 2018–2023. Animals . 2024; 14(17):2606. https://doi.org/10.3390/ani14172606

Kisley, Michael A., Esther J. Chung, and Hannah Levitt. 2024. "Investigating the Reasons behind Companion Animal Relinquishment: A Systematic Content Analysis of Shelter Records for Cats and Dogs, 2018–2023" Animals 14, no. 17: 2606. https://doi.org/10.3390/ani14172606

Article Metrics

Article access statistics, further information, mdpi initiatives, follow mdpi.

MDPI

Subscribe to receive issue release notifications and newsletters from MDPI journals

  • Open access
  • Published: 02 September 2024

Characterization of fresh biowastes for biogas production and environmental health in Niger Delta, Nigeria

  • Faith Ajiebabhio Ogbole 1 ,
  • Anthony Anwuli Ogbuta 2 &
  • Hilary Izuchukwu Okagbue 3  

BMC Environmental Science volume  1 , Article number:  8 ( 2024 ) Cite this article

Metrics details

Environmental pollution is a public health problem in Niger Delta, Nigeria. Therefore, the aims of the present study were to: identify the major fresh biowastes in Bayelsa State, Niger Delta; quantify the biogas yields from mono-digestion and co-digestion of identified biowastes; determine the first day of biogas production during hydraulic retention time; assess pH variations during anaerobic digestion; and evaluate biogas flame colours.

Fifteen communities in Bayelsa State were randomly selected, and on-the-spot assessment and quantification of the biowastes found in each community were carried out daily per week. Mono-digestion of 20 kg of each biowaste and co-digestion of 10 kg of animal waste with 10 kg of plant waste were carried out respectively under anaerobic conditions. Cumulative biogas yield and pH were measured using a pH meter and weighing scale respectively. Biogas flame colours during combustion were visually assessed.

Exactly 120.61 metric ton of fresh biowastes was found to be generated per week in Bayelsa State, Niger Delta. Industrial biowastes were the highest [47.6 tonnes, (39.46%)], followed by abattoir biowastes [33 tonnes (27%)], market and roadside sellers biowastes [25.5 tonnes (21.14%)], and farm biowastes [14.51 (12.03%)]. Biogas yields (kg) were: 0, 0, 0, 0, 0, 17, 14, 2, 18, 16, 17, 15 and 1 kg for palm oil mill effluent (POME), orange fruit waste (OFW), pineapple peels (PP), plantain peels, cassava mill effluent (CME), rumen digesta (RD), cow dung, sewage, PP-RD, plantain peels-cow dung, POME-rumen digesta, OFW-cow dung, and CME-sewage respectively. The first day of biogas production for RD, cow dung, sewage, PP-RD, plantain peels-cow dung, POME-RD, OFW-cow dung, and CME-sewage was on the 6th, 32nd. 56th, 1st, 26th, 18th, 25th, and 60th day of hydraulic retention time respectively. A dominant blue flame colour mixed reddish yellow or orange flames were found during biogas combustion. A slight increase in pH was found in all the biodigester media.

Conclusions

In the present study, a variety of biowastes yielding various quantities and qualities of biogas were identified in Bayelsa State, Niger Delta. The study’s findings have provided evidence-based data that might be explored as a road map and catalyst for policy creation against inadequate biowaste management and as sustainable alternatives to the expensive liquefied petroleum gas. The potential of current research study to be scaled up for commercial use is implicated in the present study.

Peer Review reports

Infectious diseases can be spread from biowastes , which serve as major environmental pollutants and routes for disease transmission to humans [ 1 ]. Environmental pollution is a major problem in Niger Delta, Nigeria. Indiscriminate disposal of refuse and sewage around Bayelsa State capital, Yenagoa, and Niger Delta environs, as well as in the mangrove forest and water bodies in Niger Delta is a regular occurrence [ 2 , 3 ]. Open dump sites with foul smells are found along major roads in Niger Delta [ 4 ]. These call for the need for effective environmental pollution management system in Niger Delta. The recycling of biowastes into useful biogas might serve as an alternative to the current inadequate management of environmental pollution in Niger Delta and Nigeria as a whole. It might also serve as an alternative to the expensive liquefied petroleum gas (cooking gas) in Niger Delta and Nigeria at large. According to the National Bureau of Statistics (NBS), Bayelsa State in Niger Delta had the highest retail price for LPG (cooking gas) in 2023, although the state is a major producer of LPG [ 5 ]. Also, according to NBS, since the beginning of 2024 till date, the average retail price of LPG in Nigeria has been increasing by 28.33% on a month-on-month basis and by 46.88% on a year-on-year basis from 2023 to 2024 [ 6 ]. The ever-increasing cost of liquefied petroleum gas (LPG) in Nigeria calls for the need to research on cheaper alternatives to LPG.

A previous report by the International Renewable Energy Agency in 2020 showed that the cost of biogas is 50% less than the cost of LPG. As a result, waste-to-energy technology involving anaerobic digestion of biowastes for biogas production is emerging as an alternative to LPG in other continents and countries [ 7 ], with Europe, China, and the United States accounting for 90% of global biogas production [ 8 ] and India practicing household biogas production [ 9 ]. Nigeria and Africa at large are yet to fully embrace biogas production on a national or continental scale.

The Niger Delta region of Nigeria possesses a unique climate, culture, and tropical vegetation that influences their way of life, occupation, available industry, and the type of waste they generate [ 10 , 11 ]. From one nation to another, the quantity of waste and the frequency at which it is produced vary. Even within a given nation , the quantity of waste, accessibility, composition, and frequency at which it is produced also vary due to several factors, such as the type of industries present, the human activities taking place, the level of affluence, customs, climate, and living conditions [ 12 , 13 ].

A study carried out in Lyon, a city in France, found that garden biowastes, restaurant biowastes, household biowastes, and supermarket biowastes were the major biowastes found for biogas production [ 14 ]. Another study carried out in India found that grass, hemp, wheat straw, leaves, cattle manure, pig manure, poultry manure, peanut hull, yard waste, maize, cattle slurry, sewage sludge, rice straw, municipal solid waste, wastewater leachate, pulp and paper mill sludge, food waste, and kitchen waste were the major biowastes found for biogas production [ 15 ]. In Malaysia, animal dung was found to be the major biowaste for biogas energy production [ 16 ]. Various forms of biowastes have also been found in other geographical locations [ 17 , 18 ]. Thus, characterizing the biowastes in a given locality may help determine the feasibility of sustainable biogas production in that area. It may also help to determine the most suitable biogas production methodology to be employed [ 17 , 18 ].

Aside from serving as biowastes, animal faeces, such as cow dung, also serve as inoculum for biogas production because they contain all the microorganisms, especially methanogens, needed for the anaerobic conversion of biowastes to biogas [ 19 , 20 ]. The co-digestion of plant wastes and animal wastes has been reported to enhance biogas production [ 19 , 20 ]. This justifies the co-digestion of plant waste and animal faeces. Although the justification for the co-digestion of plant wastes and animal faeces is well known, evidence-based data to support the absence of biogas production from the mono-digestion of plant wastes is sparse in the literature. This gap in research is what the present study intends to fill.

The core of Niger Delta, Nigeria, is Bayelsa State. The state is inhabited by the Ijaw-speaking tribe of Nigeria, and their major occupations are fishing, farming, palm oil milling, lumbering, local gin making, trading, and carving [ 21 , 22 ]. Poor waste management and environmental pollution has been previously reported in this region [ 2 , 3 , 4 ]. However, studies on waste management and environmental pollution management through the recycling of biowastes for biogas production have not been previously reported in Bayelsa State, Niger Delta. Information from such studies are needed to inform the feasibility of sustainable biogas production in the State.

Therefore, the aims of the present study were to: identify the major fresh biowastes in Bayelsa State, Niger Delta; quantify the biogas yields from mono-digestion and co-digestion of identified biowastes; determine the first day of biogas production during hydraulic retention time; assess pH variations during anaerobic digestion; and evaluate biogas flame colours.

Study location

The study was conducted in Bayelsa State, Niger Delta. Bayelsa State is one of the 36 states in Nigeria and is situated in the core of Niger Delta. The coordinates of Bayelsa State are 6.06990° E latitude and 4.77190° N longitude. The State primarily consists of rural settlements, except for the capital city, which features urban areas. This rural predominance can be attributed to the riverine and swampy nature of the terrain, which hinders major infrastructural development. Bayelsa State is a coastal State that is heavily waterlogged with a high underground water table. Many of the indigenous people live by the seashore, or along river banks and engage in fishing and farming as their primary occupations [ 23 , 24 ]. Bayelsa State comprises eight Local Government Areas (LGAs) which are Ekeremor, Kolokuma/Opokuma, Yenegoa (the state capital), Nembe, Ogbia, Sagbama, Brass, and Southern Ijaw, as shown Fig.  1 .

figure 1

Maps of Nigeria and Bayelsa State. The map of Nigeria shows the location of Bayelsa State within Nigeria while the map of Bayelsa State shows the eight LGAs in the State [ 25 ]

Sampling of the fifteen representative study communities in Bayelsa State

Multistage random sampling was used to select fifteen representative communities for the study [ 26 ]. In the first sampling stage, six LGAs were randomly selected out of the eight LGAs in the state, as shown in Table  1 . In the second stage, two communities were randomly selected from each of the selected LGAs, except for Yenegoa, where four communities were randomly selected, and Sagbama LGA, where three communities were randomly selected, as shown in Table  1 . Four communities were randomly selected from Yenegoa because Yenegoa is the state capital city and has several markets as well as roadside vendors that generate biowastes. Three communities were randomly selected from Sagbama LGA because Sagbama LGA houses three tertiary institutions and several agro-allied industries that generate biowastes.

Identification and quantification of major biowastes in Bayelsa State, Niger Delta, Nigeria

On-the-spot assessment of the biowastes in each of the selected communities was carried out on a daily basis per week per community. For each community, a tour around the community to map the available biowaste clusters and roadside sellers that generate biowastes was first carried out. Then the types and quantities of biowastes generated on a daily basis were noted and quantified using a digital weighing scale for a period of one week for each community. For large quantities of biowastes, the total area or volume where the biowaste was found was measured and recorded. Then a small area was marked out, and the biowaste in that small area was collected, categorized by sorting, and each category was quantified using a weighing scale. The quantity of biowaste in the small area was used to extrapolate the total quantity of biowaste in the whole area. For locations where the biowastes were scattered around, the biowastes were first gathered together before measurement and quantification. A representative sample of each major biowaste was collected from each study location for biogas production. The recorded weights and volumes were used for data analysis, and the results were expressed as tonnes of biowastes per week [ 14 ]. The handling of biowastes was carried out according to the National Environmental Standards and Regulations Enforcement Agency’s (NESREA) guidelines for the handling of environmental wastes [ 27 ].

Anaerobic biodigester fabrication and set up

Polyvinyl chloride (PVC) drums were fabricated into biodigesters. One of the fabricated biodigesters is shown in Fig.  2 a. Briefly, three openings were made on each PVC drum. The first opening, a large hole, was created on top of the drum and was fitted with a biowaste inlet pipe. A second opening (a smaller hole) was created on top of the PVC drum and was fitted with a small biogas outlet pipe. The third opening was created at the side of the PVC drum and was fitted with an effluent pipe. A control valve was fitted on each pipe. The biogas outlet pipe was connected to a biogas storage bag through a hose. A representative biogas storage bag made from tarpaulin and fitted with a hose was made for biogas storage as shown in Fig.  2 b. Each set-up was made airtight by applying silicon gum to the edges of the pipes and bags [ 14 ]. Typical operation pressures for biogas storage balloon ranging from 2—8 mBar [ 28 ] was used in the present study. The pressure used for the present study was set at 4 mBar [ 28 ].

figure 2

a A representative biodigester set up and b A representative biogas storage bag

The inlet pipe is the large pipe on top of the drum used to feed biowastes (substrates) into the anaerobic biodigester. The effluent pipe is the pipe at the side of the drum used for the discharge of liquid waste out of the anaerobic biodigester. The biogas outlet pipe is the smaller pipe on top of the drum used for the passage of biogas from the anaerobic biodigester into the biogas storage bag through a hose. The storage bag (Fig.  2 b) is flat, indicating that no biogas has been produced. The biodigester was placed under direct sunlight to warm it up to its optimal anaerobic temperature, while the biogas storage bag was kept inside a store room.

Moisture content determination of biowastes

Loss of water during oven drying was used to determine moisture content and the result was expressed in percentage. Briefly, 5 kg of each sample was placed in an oven for 16 h at 105 ± 2 °C and dried until a constant weighed was obtained. The moisture content was calculated by using the formular below [ 29 ].

Mono- and co-anaerobic digestion of identified biowastes

Given that previous studies have shown that the mixing ratio of animal faeces and organic waste is 1:1 and that this ratio ensures that there is a balanced mixture of carbon and nitrogen, which is essential for the efficient breakdown of organic matter and the production of biogas [ 19 , 20 ], the present study utilized a mixing ratio of 1:1 for animal faeces and plant wastes for co-digestion. Also, the widely used mixing ratio of 1:1 for animal faeces and water [ 30 ] was utilized in the present study for mono-substrate digestion. Solid plant waste samples were grinded and made into slurry by adding equal volume of water before feeding into the anaerobic biodigester. After feeding, the first day of biogas production during hydraulic retention time was observed and recorded [ 19 , 20 , 30 ].

Flame colour test

Visual examination of the colours of the flames produced during the combustion of the various biogas was carried out as previously described [ 31 , 32 ].

Measurement of biogas yield

A weighing scale in kilograms was used to measure the biogas produced daily for one week and results were recorded as cumulative biogas yield per week [ 14 , 19 ].

Measurement of the pH of the biodigester medium

The initial and final pH of the biodigester medium were measured using a using glass electrode pH meter with a ratio of 2:1 for biowastes to water suspension [ 19 ].

Statistical analysis

SPSS version 22 was used for data analysis. Values were presented as cumulative sums and as mean ± standard error of mean.

Major fresh biowastes in Bayelsa State, Niger Delta, Nigeria

Results for the major fresh biowastes found in Bayelsa State, Niger Delta, are presented in Table  2 and Fig.  3 , respectively. As shown in Fig.  3 , a variety of biowastes were found. Also, as shown in Table  2 , 120.61 metric ton of fresh biowastes was found to be generated per week in Bayelsa State, Nigeria. Industrial biowaste was the highest group of biowaste [47.6 tonnes, (39.46%)], followed by abattoir biowaste [33 tonnes (27%)], market and roadside sellers biowastes [25.5 tonnes (21.14%)], and farm biowastes [14.51 (12.03%)]. The most common biowastes found in all the study locations were plantain peels and empty fruit bunches of plantain. A significant difference was found in the biowastes generated from the selected communities. Cassava mill effluent was found to be channeled into nearby rivers and creeks in most cassava mills. Thus only a small quantity of cassava effluent could be assessed and collected for quantification. Additional information on Table 2 is presented as a Supplementary material.

figure 3

A cross section of some biowastes found in the present study during on-the-spot assessment of biowastes in Bayelsa State, Niger Delta, Nigeria. a Empty bunches of oil palm fruits, b Palm oil mill effluent inside a well, c Empty fruit bunches of plantain, d A bucket of cow dung slurry, e Pineapple peels, f Cassava peels, g Sheaths of corn

Moisture state of the feedstock selected for biogas production and the state at which the material was grinded

The moisture state of the feedstock selected for biogas production and the state at which the materials were grinded are presented in Table  3 . POME, cassava mill waste water, and sewage from septic tanks were in a liquid state. These required no grinding and no further addition of water before feeding into the biodigester. The sewage used in the present study was very watery with a very high moisture content. Excessive seepage of underground water into septic tanks was found in the present study.

Biogas yield from mono-digestion and co-digestion of biowastes used as feedstock

Figures  4 and 5 show the cumulative biogas yield from the mono-digestion and co-digestion of the biowastes used as feedstock for biogas production in the present study. Beginning from the first day that biogas production was observed, biogas was collected daily from the biodigester over a period of 7 days. Thus Figs.  4 and 5 show the cumulative biogas yield over a period of 7 days. As shown in Fig.  4 , mono-digestion of animal waste yielded biogas as well as the co-digestion of plant waste and animal waste. However, mono-digestion of plant wastes yielded no biogas. Co-digestion of pineapple peels and rumen digesta yielded the highest quantity of biogas (18 kg of biogas/20 kg of biowastes), while co-digestion of cassava mill waste water with sewage yielded the lowest quantity of biogas (1 kg of biogas/20 kg of biowastes). However, plant-only digestion yielded carbon dioxide gas that extinguished a glowing splinter and turned lime water milky. Also, as shown in Fig.  5 , inflated storage bags signify the presence of biogas.

figure 4

Cumulative biogas yield collected over a period of 7 days from the mono-digestion and co-digestion of 20 kg of biowastes. POME: Palm oil mill effluent

figure 5

A representative cross section of the different biogas storage bags used for the daily collection of the biogas generated from the mono-digestion and co-digestion of biowastes in Bayelsa State, Niger Delta, Nigeria. Black tyre tube, 2 L (left), small biogas storage bag, 500 L (left), medium biogas storage bag, 4000 L (right). The maximum pressure the bags can withstand 4 mBar

Comparison of the beginning of biogas production for the various feedstock during hydraulic retention time

Figure  6 shows the variations that were found in biogas production commencement date during hydraulic retention time among the various biowastes. Figure  6 shows the first day that biogas emission from the anaerobic biodigester was observed after feeding and not the duration of complete digestion. Production of biogas started on the 6th, 32nd. 56th, 1st, 26th, 18th, 25th, 60th day of the retention period for rumen digester (RD), cow dung, sewage, pineapple peels (PP) and RD mixture, plantain peels and cow dung mixture, POME and RD mixture, orange fruit wastes (OFW) and cow dung mixture, and cassava mill effluent (CME) and sewage mixture respectively. Co-digested pineapple peels and rumen digesta had the earliest biogas production commencement date, while co-digested cassava wastewater (cassava mill effluent) and human faeces had the longest biogas production commencement date during hydraulic retention time.

figure 6

First day of biogas production for various biowastes during hydraulic retention time in the biodigester. Values for the mono-digestion of plant substrates not available because no biogas was produced from the mono-digestion of plant wastes

Flame colours of the various biogas produced

Table 4 shows the colours of the flames generated during the combustion of the various biogas produced in the present study. As shown in Table  4 , a dominant blue flame was visualized in all the biogas produced, which was mixed with various shades of reddish yellow or orange flames.

Variations in pH during anaerobic digestion of biowastes

As shown in Fig.  7 , an increase in pH during anaerobic digestion of biowastes was observed across all the biodigester media. Slight variations between the mean initial and final pH were found. The pH values for the media that produced combustible biogas varied from low acidity to slightly alkaline, while the pH values for the plant-based biowastes that yielded no biogas varied within the acidic range.

figure 7

In the present study, we have characterized the biowastes found in Bayelsa State, Niger Delta. The quantity of biowastes and locations where they can be abundantly found on a weekly basis within Bayelsa State, Niger Delta, have been presented in this study. The biogas quality in terms of flame colours and the first day of biogas production during hydraulic retention time in the biodigester, as well as the quantity of biogas produced from the mono-digestion and co-digestion of substrates (biowastes), have also been presented in this study. Finally, variations in pH during the anaerobic digestion of the different biowastes were also determined in the present study.

Domestic biowastes were not found as major biowastes in the present study. On the contrary, a previous study carried out in Lyon, a city in France, found that domestic biowastes were a major component of biowastes [ 14 ]. Also, although another study carried out in Central African Rebulic found that market biowaste was not a major component of biowastes [ 17 ], the present study found that market and road seller`s biowastes (empty bunches of plantain peels, pineapple peels, and orange peels) were major components of biowastes. This shows that from one region of the world to another, the composition of biowastes varies. Thus, in carrying out biogas production, a knowledge of the biowastes locally available as raw materials or feedstock in a given area is important.

Also, although human faeces have been previously reported as a good raw material for biogas production [ 33 ], however, the present study found a low quantity of biogas yield from human faeces (sewage) as shown in Fig.  1 . This might be due to the high moisture content of the sewage used in the present study. The high moisture content was due to the excessive watery nature of the sewage. The excessive watery nature of the sewage was as a result the seepage of underground water into the sewage in the septic tank, thereby resulting in the excessive watery sewage utilized in the present study. Bayelsa State lies on a swampy terrain with a high underground water table [ 34 ]. The present study found that seepage of underground water into septic tanks in Bayelsa State occurred in most septic tanks visited in the communities selected for the present study [ 34 ]. This excessively watery sewage together with the excessively watery cassava mill waste water used in the present study might be responsible for the decrease in biogas yield of sewage upon co-digestion with cassava mill waste water which was found in the present study.

Studies comparing the biogas yield of several biowastes under the same set of experimental study conditions are sparse in literature [ 19 , 20 , 29 , 32 , 33 ]. The present study investigated the feasibility of biogas production from industrial biowastes, market and roadside sellers’ biowastes, farm biowastes, and abattoir biowastes. This has provided a snapshot of the contributions of several biowastes to biogas production. Mono-digestion of plant biowastes did not yield biogas in the present study. Thus, this pattern of biowaste digestion is not recommended for future biogas production. When producers are provided with the biogas production efficiency of several locally available biowastes, they will be able to make a better informed choice of the best raw material to use for biogas production, as well as the biogas production process to be employed. This will make biogas production more flexible, feasible, sustainable, and appealing.

Hydraulic retention time is the duration that feedstock stay in a digester during anaerobic digestion. Biogas production can commence any day within the hydraulic retention time and continue till the last day of the retention time [ 35 , 36 , 37 , 38 ]. “ Comparison of the beginning of biogas production for the various feedstock during hydraulic retention time ” section, showed the biogas production commencement dates for the various biowastes during hydraulic retention time. Compared with other biowastes, the co-digestion of pineapple peels and rumen digesta had the earliest biogas production commencement date during hydraulic retention period. This might be due to the variations in the proximate composition of the various biowastes as previously reported [ 35 , 36 , 37 , 38 ]. The high content of simple sugars and other simple carbon compounds in pineapple peels, may have made pineapple peels very easy to digest [ 35 ]. This may have resulted in the early biogas production commencement observed for pineapple peels mixed with rumen digesta unlike the other biowastes which had high content of complex carbon compounds such as cellulose and lignin which take longer periods to be broken down [ 39 ]. the present study found that the complex biowastes had longer commencement date for biogas production during hydraulic retention time or digestion period [ 39 ]. The findings of the present study corroborate the findings of previous studies [ 36 , 37 , 38 ]. Thus, a careful selection of biowastes with early biogas production start date may improve biogas production efficiency. Characterization of the commencement date for biogas production of a variety of biowastes during hydraulic retention time is limited in literature. Retention times are usually given in literature without indication of the exact biogas production commencement date within the retention time. Thus, the present study has filled a major gap in research. It must be noted that the values presented in Fig.  6 referred to the first day that biogas emission from the biodigester was observed after feeding the digester and not the day that complete digestion of the biowastes occurred. For example, although biogas emission was observed within 24 h from the biodigester containing a mixture of pineapple peels and rumen digesta however, complete digestion continued till the next seven days, as evident by the continuous production of biogas till day seven of biogas emission [ 35 , 36 , 37 , 38 ].

The present study also found that the various biogas produced in the present study gave a dominant blue flame colour and various shades of orange or reddish yellow flames during the combustion. Flame colours of biogas have been previously reported in a study carried out by Ketuk et al .  [ 31 ], where the flame color of the biogas from cow dung and goat manure starters was found to be redder than the flame colour from the biogas from leachate starter. Also, another previous study found that the average flame colour of unpurified biogas from coffee waste materials was 60.16% blue and 39.84% red flame [ 32 ]. However, the present study has expanded the findings of theses previous studies [ 31 , 32 ] by providing information on the flame colours produced of the biogas produced from a range of other substrates not previously reported. Previous studies have showed that biogas with a high degree of reddish yellow or orange flames are of lower quality and generates lower heating temperature compared with biogas with blue flame which are of higher quality and generates higher heating temperature [ 31 , 32 ]. In the present study the biogas from POME was redder than the biogas produced from the other biowastes. This implies that heating with the biogas from POME will take a longer time due to its lower temperature compared with biogas with a higher degree of blue flame. The findings of the present study is quite novel given that till date, no study has characterized the flame characteristics of the biogas produced from a range of biowastes in a single study. The present study has thus helped to fill the gap in research.

Flame colour is a qualitative measure of biogas purity. Flame testing can be used to determine fuel quality because flame is a combustion chemical reaction and burning is influenced by fuel quality [ 31 , 32 ]. A yellow or red flame on gas stove is dangerous, as it is indicative of incomplete combustion and carbon monoxide (CO) generation [ 31 , 32 ]. This could lead to health hazard and wastage of biogas as well as taking longer period of time for heating to occur. It can also result in the generation of soot on the appliances used for heating [ 31 , 32 ]. In addition, the present study found that plant only biowastes did not generate any flame, rather they emitted a gas that extinguished a glowing splinter. This suggests that plant-based wastes must be co-digested with animal intestinal or feacal waste for biogas production to occur. However, further study is recommended.

It is a known fact that pH is an important factor in biogas production [ 38 ]. Non-methanogenic bacteria, which carry out the first stage of anaerobic digestion, can adapt to a pH range of 4.0–8.5, while methanogenic bacteria, which carry out methanogenesis, the second stage of anaerobic digestion, are limited to a pH range of 6.5–7.2 [ 38 ]. This might explain why only the mediums with a nearly neutral to slightly alkaline pH produced more biogas in the present study compared with mediums with a lower pH.

The underlying uniqueness and significance of the study’s findings, particularly in terms of biowaste management in the Niger Delta, is that the study`s findings have provided evidence-based data that might be explored as a road map for policy creation against the inadequate biowaste management, and environmental pollution in Niger Delta [ 2 , 3 , 4 ]. In addition, the underlying uniqueness and significance of the study’s findings, particularly in terms of biogas production in Niger Delta, is that the study`s findings have provided an action plan that might serve as a catalyst for the implementation of biogas production using the available biowastes in Niger Delta as cheap sources of feedstock. This might serve as an alternative to the highly expensive cooking gas in the Niger Delta and Nigeria as a whole [ 5 , 6 ].

The present study highlighted the qualitative and quantitative characteristics of the major biowastes found in Bayelsa State, Niger Delta. The present study also demonstrated variations in the yield and flame colours of the biogas produced from various biowastes. Variations in the commencement date for biogas production during anaerobic digestion hydraulic retention time for the various biowastes were also demonstrated in the present study. The present study also found excessive moisture content in the sewage from septic tanks in Bayelsa State due to the high underground water table in Bayelsa State. Further study to dewater the sewage in Bayelsa State, Niger Delta, for efficient biogas production is recommended. The potential of current research study to be scaled up for commercial use is implicated in the present study.

Availability of data and materials

The data is not publicly available but can be provided upon reasonable request from the corresponding author. Moreover, some are provided as Supplementary data.

Abbreviations

Carbon monoxide

Empty fruit bunch of oil palm

Liquefied petroleum gas

National Bureau of Statistics

Palm oil mill effluent

Pineapple peels + Rumen digesta

Polyvinyl chloride

Faiza NMT, Hassan NA, Farhan MR, Edre MA, Rus RM. Solid waste: Its implication for health and risk of vector borne diseases. JWBM. 2019;1(2):14–7.

Google Scholar  

Aroloye O. Numbere AO. Municipal solid waste disposal in mangrove forest: environmental implication and management strategies in the Niger Delta, Nigeria. Municip Solid Waste Manag. 2019;Ch. 10. https://doi.org/10.5772/intechopen.83809 .

Okpara DA, Kharlamova M, Grachev V. Proliferation of household waste irregular dumpsites in Niger Delta region (Nigeria): unsustainable public health monitoring and future restitution. Sustain Environ Res. 2021;31:4. https://doi.org/10.1186/s42834-020-00077-1 .

Article   CAS   Google Scholar  

Ebuete AW, Wodu DP, Ebuete E. Dumping on waters: the lacunae in waste management in the Niger Delta, Nigeria. AJEC. 2022;1(2):100–9.

National Bureau of Statistics. Liquefied Petroleum Gas (Cooking Gas) Price Watch. 2023. https://nigerianstat.gov.ng/elibrary/read/1241345 . Accessed May 2023.

National Bureau of Statistics. Liquefied Petroleum Gas (Cooking Gas) Price Watch. 2024. https://nigerianstat.gov.ng/elibrary/read/1241476 . Accessed Feb 2024.

International Renewable Energy Agency Renewable Capacity Statistics. International Renewable Energy Agency: Masdar City; 2020.  https://www.irena.org/publications/2020/Mar/Renewable-Capacity-Statistics-2020 .

IEA. Outlook for biogas and biomethane: prospects for organic growth. Paris: IEA; 2020. https://www.iea.org/reports/outlook-for-biogas-and-biomethane-prospects-for-organic-growth . Accessed March 2020.

Mittal S, Ahlgren EO, Shukla PR. Barriers to biogas dissemination in India: a review. Energy Policy. 2018;112:361–70.

Ogbole FA, Igwe CU, Onuoha HC, Nzebude CP. Characterization of ABO / Rhesus Antigen Polymorphism Associated with Malaria in a Malaria Hotspot in Bayelsa State, Niger Delta, Nigeria. IJBRR. 2023;32(7):42–52.

Ogbole FA, Oyelana O. Health Risk Assessment of the Drinking Water from Sagbama River, Bayelsa State, Niger Delta, Nigeria. IJSER. 2020;11(9):1455–60.

Maalouf A, Mavropoulos A. Re-assessing global municipal solid waste generation. Waste Manag Res. 2023;41(4):936–47.

Article   Google Scholar  

Noufa M, Yuanyuan L, Maalla Z, Adipah S. Determinants of household solid waste generation and composition in Homs City, Syria. JESPH. 2020. https://doi.org/10.1155/2020/7460356 .

Moretti P, Morais de Araujo J, Borges de Castilhos A, Buffière P, Gourdon RB. Characterization of municipal biowaste categories for their capacity to be converted into a feedstock aqueous slurry to produce methane by anaerobic digestion. Sci Total Environ. 2020;716:1–18.

Bhatia RK, Ramadoss G, Jain AK, Dhiman RK, Bhatia SK, Bhatt AK. Conversion of waste biomass into gaseous fuel: present status and challenges in India. Bioenerg Res. 2020;13:1046–68. https://doi.org/10.1007/s12155-020-10137-4 .

Atchike DW, Irfan M, Ahmad M, Rehman MA. Waste-to-renewable energy transition: biogas generation for sustainable development. Front Environ Sci. 2022. https://doi.org/10.3389/fenvs.2022.840588 .

Dima FAFJ, Li Z, Mang HP, Zhu L. Feasibility analysis of biogas production by using GIS and multicriteria decision aid methods in the central African Republic. Sustainability. 2022;14(20):13418. https://doi.org/10.3390/su142013418 .

Ebunilo P, Aliu S, Orhorhoro K. Comparative analysis of biogas yield from different composition of domestic wastes from Benin City, Nigeria. JAAS. 2015;4(5):169–77.

Chibueze U, Okorie N, Oriaku O, Isu J, Peters E. The production of biogas using cow dung and food waste. J Mater Chem. 2017;7(2):21–4.

CAS   Google Scholar  

Kamble R, Ingale S. Performance study of biogas plant using kitchen and food waste. IJAERS. 2018;5(1):120–6.

Ogbole FA. Urinalysis for dehydration, kidney injury and urinary tract infection assessment in uural Riverside, Bayelsa State, Nigeria. IJISRT. 2021;6(12):805–10.

Ogbole FA, Harold BA. Association of undiagnosed pre-diabetes and type-2 diabetes mellitus with interleukin-2 mRNA expression among adults in Bayelsa State, Nigeria. IJRR. 2023;10(5):210–5.

Ogbole FA, Moroyei BE. Upregulation of interleukin-2 among hypertensive subjects in Bayelsa State, Nigeria. IJRR. 2023;10(6):557–65.

Ogbole FA, Sanugba EE. Distribution of body mass index and abdominal obesity in Bayelsa State with associated interleukin-2 gene expression. AJBGMB. 2023;14(4):1–10.

DailyTrust Flood: Bayelsa SEMA Confirms 96 Deaths, Over 1.2m Displaced Persons. Available at: https://dailytrust.com/flood-bayelsa-sema-confirms-96-deaths-over-1-2m-displaced-persons/ . Last accessed: 7 Nov 2022

Ogbole FA, Fatusi AO. Co-occurrence of health risk behaviours for communicable and non-communicable diseases in a tertiary institution in South West Nigeria. IJRSI. 2020;7(1):10–4.

National Environmental Standards and Regulations Enforcement Agency (NESREA). Laws and Regulations. https://www.nesrea.gov.ng/publications-downloads/laws-regulations/ . Accessed 11 Feb 2024.

Kube J. Management of gas storages in biogas plants. Chem Eng Technol. 2018;41:702–10. https://doi.org/10.1002/ceat.201700395 .

Chryssou K, Stassinopoulou M, Lampi, E. Calculation of the expanded uncertainty of the measurements of the moisture content of a lot of paper samples oven dried at 105 0 C. Res Rev. 2018. https://www.rroij.com/open-access/calculation-of-the-expanded-uncertainty-of-the-measurements-of-the-moisture-content-of-a-lot-of-paper-samples-oven-dried-at-105c.php?aid=87181 .

Benali M, Hamad T, Hamad Y. Experimental study of biogas production from cow dung as an alternative for fossil fuels. J Sustain Bioenergy Syst. 2019;7:1–13.

Ketut CN, Agung S, Mekro P, Heri HB. The flame characteristics of the biogas has produced through the digester method with various starters. IOP Conf Ser: Mater Sci Eng. 2018;299:012091. https://doi.org/10.1088/1757-899X/299/1/012091 .

Ilminnafik N, Setiyawan DL, Sutjahjono H, Rofiq A, Hadi AS. Flame characteristics of biogas from coffee waste materials. IOP Conf. Series: J Phys Conf Ser. 2019;1175. https://doi.org/10.1088/1742-6596/1175/1/012273 .

Donacho DO, Tucho GT, Olani DD, Kabtiyimer HE, Hailu AB, Wolde AD. Experimental evaluation of fresh human feces biogas and compost potential: evidence for circular economy from waste streams in Ethiopia. Heliyon. 2023;9(12):e22494.

Ogbole FA, Igwe CU, Onuoha HC, Nzebude CP. Evaluating the prevalence of malaria parasite infection among adults in wetlands using nested PCR and high resolution melting analysis. AJBGMB. 2023;14(4):53–63.

Siti RAM, Zainal SNA, Nadzirah KZ. Determination of sugar content in pineapple waste variety N36. Int Food Res J. 2013;20(4):1941–3.

Tasnim A, Al Mamun MR, Hossen MA, Rahman MT, Soeb MJA. Comparison effect on biogas production from vegetable and fruit waste with rumen digesta through co-digestion process. EJ-Energy. 2022;2(1):1–7.

Ogbole FA, Akemi CO. GC-MS analysis of biogas from pineapple peels and toxicological evaluation of generated effluent. AJBGE. 2023;6(2):96–104.

Zhang C, Su H, Baeyens J, Tan T. Reviewing the anaerobic digestion of food waste for biogas production. Renew Sustain Energy Rev. 2014;38:383–92.

Deng C, Kang X, Lin R, Wu B, Ning X, Wall D, Murphy JD. Boosting biogas production from recalcitrant lignin-based feedstock by adding lignin-derived carbonaceous materials within the anaerobic digestion process. Energy. 2023;278:127819.

Download references

Acknowledgements

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Author information

Authors and affiliations.

Biochemistry Unit, Department of Chemical Sciences, University of Africa Toru – Orua, Toru-Orua, Bayelsa State, Nigeria

Faith Ajiebabhio Ogbole

Chemistry Unit, Department of Chemical Sciences, University of Africa Toru – Orua, Toru-Orua, Bayelsa State, Nigeria

Anthony Anwuli Ogbuta

Sydani Institute for Research and Innovation, Sydani Group, Abuja, Nigeria

Hilary Izuchukwu Okagbue

You can also search for this author in PubMed   Google Scholar

Contributions

FAO, AAO and HIO designed the concept and scope of the work. FAO, AAO and HIO carried out field survey for data collection. FAO and HIO analyzed and interpreted the data regarding the quantity of fresh biowastes, biogas yield and pH. FAO, AAO and HIO reviewed the manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Hilary Izuchukwu Okagbue .

Ethics declarations

Ethics approval and consent to participate.

The study does not involve human participants. Hence, no ethical approval was needed.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s note.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Supplementary material 1, rights and permissions.

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Cite this article.

Ogbole, F.A., Ogbuta, A.A. & Okagbue, H.I. Characterization of fresh biowastes for biogas production and environmental health in Niger Delta, Nigeria. BMC Environ Sci 1 , 8 (2024). https://doi.org/10.1186/s44329-024-00008-0

Download citation

Received : 23 April 2024

Accepted : 12 August 2024

Published : 02 September 2024

DOI : https://doi.org/10.1186/s44329-024-00008-0

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Mono-substrate digestion
  • Co-substrate digestion
  • First day of biogas production

BMC Environmental Science

ISSN: 3004-8710

randomly selected in research

IMAGES

  1. Simple Random Sampling: 6 Basic Steps With Examples

    randomly selected in research

  2. Random Assignment in Experiments

    randomly selected in research

  3. Randomisatie in wetenschappelijke experimenten

    randomly selected in research

  4. How Random Selection Is Used In Research

    randomly selected in research

  5. An Overview of Simple Random Sampling (SRS)

    randomly selected in research

  6. Randomly selected experimental groups.

    randomly selected in research

VIDEO

  1. WHO ARE THE RESPONDENTS OF YOUR STUDY?

  2. What is Research methodology? definition and it's types

  3. Dominik Beck: Mean distance in polyhedra, an application of the Crofton Reduction Technique

  4. Handsome Old Man

  5. Quantitative Researcher Interviews

  6. Sampling

COMMENTS

  1. How Random Selection Is Used For Research

    The Random Selection Experiment Method. When researchers need to select a representative sample from a larger population, they often utilize a method known as random selection. In this selection process, each member of a group stands an equal chance of being chosen as a participant in the study.

  2. Simple Random Sampling

    Revised on December 18, 2023. A simple random sample is a randomly selected subset of a population. In this sampling method, each member of the population has an exactly equal chance of being selected. This method is the most straightforward of all the probability sampling methods, since it only involves a single random selection and requires ...

  3. Sampling: how to select participants in my research study?

    The essential topics related to the selection of participants for a health research are: 1) whether to work with samples or include the whole reference population in the study (census); 2) the sample basis; 3) the sampling process and 4) the potential effects nonrespondents might have on study results. We will refer to each of these aspects ...

  4. What Is Probability Sampling?

    What Is Probability Sampling? | Types & Examples

  5. What Is Random Selection?

    Dictionary. Random selection refers to a process that researchers use to pick participants for a study. When using this method, every single member of a population has an equal chance of being chosen as a subject. This process is an important research tool used in psychology research, allowing scientists to create representative samples from ...

  6. Sampling Methods

    When you conduct research about a group of people, it's rarely possible to collect data from every person in that group. ... From the first 10 numbers, you randomly select a starting point: number 6. From number 6 onwards, every 10th person on the list is selected (6, 16, 26, 36, and so on), and you end up with a sample of 100 people. ...

  7. Elements of Research : Random Selection

    Random selection. Random selection is a form of sampling where a representative group of research participants is selected from a larger group by chance. This can be done by identifying all of the possible candidates for study participation (e.g., people attending the County fair on a Tuesday) and randomly choosing a subset to participate (e.g ...

  8. Randomization

    Randomization is a statistical process in which a random mechanism is employed to select a sample from a population or assign subjects to different groups. [1] [2] [3] The process is crucial in ensuring the random allocation of experimental units or treatment protocols, thereby minimizing selection bias and enhancing the statistical validity. [4]It facilitates the objective comparison of ...

  9. Principles and methods of randomization in research

    Randomization is an important technique in research because, when accomplished successfully, it not only removes potential personal bias from research but also removes variables from the analysis that might confound the results. ... Blocked randomization with randomly selected block sizes. Int J Environ Res Publ Health, 8 (1) (2011), pp. 15-20 ...

  10. 15.1: Random Selection

    Introduction. The usual treatments deal with a single random variable or a fixed, finite number of random variables, considered jointly. However, there are many common applications in which we select at random a member of a class of random variables and observe its value, or select a random number of random variables and obtain some function of those selected.

  11. Sampling Methods & Strategies 101 (With Examples)

    Simple random sampling. Simple random sampling involves selecting participants in a completely random fashion, where each participant has an equal chance of being selected.Basically, this sampling method is the equivalent of pulling names out of a hat, except that you can do it digitally.For example, if you had a list of 500 people, you could use a random number generator to draw a list of 50 ...

  12. Random Selection vs. Random Assignment

    They recruit 100 individuals to be in the study by using a computer to randomly select 100 names from a database. Once they have the 100 individuals, they once again use a computer to randomly assign 50 of the individuals to a control group (e.g. stick with their standard diet) and 50 individuals to a treatment group (e.g. follow the new diet). ...

  13. Difference between Random Selection and Random Assignment

    Random selection is thus essential to external validity, or the extent to which the researcher can use the results of the study to generalize to the larger population. Random assignment is central to internal validity, which allows the researcher to make causal claims about the effect of the treatment. Nonrandom assignment often leads to non ...

  14. Random Selection & Assignment

    Random selection is how you draw the sample of people for your study from a population. Random assignment is how you assign the sample that you draw to different groups or treatments in your study. It is possible to have both random selection and assignment in a study. Let's say you drew a random sample of 100 clients from a population list ...

  15. Sampling Methods In Reseach: Types, Techniques, & Examples

    Sampling Methods In Reseach: Types, Techniques, & ...

  16. Random Assignment in Psychology: Definition & Examples

    Olivia Guy-Evans, MSc. In psychology, random assignment refers to the practice of allocating participants to different experimental groups in a study in a completely unbiased way, ensuring each participant has an equal chance of being assigned to any group. In experimental research, random assignment, or random placement, organizes participants ...

  17. 5.5: Importance of randomization

    Introduction. If the goal of the research is to make general, evidenced-based statements about causes of disease or other conditions of concern to the researcher, then how the subjects are selected for study directly impacts our ability to make generalizable conclusions.The most important concept to learn about inference in statistical science is that your sample of subjects, upon which all ...

  18. Random Assignment in Experiments

    Random Assignment in Experiments | Introduction & ...

  19. Sampling Methods

    The sample should be selected randomly, or if using a non-random method, every effort should be made to minimize bias and ensure that the sample is representative of the population. Collect data: Once the sample has been selected, collect data from each member of the sample using appropriate research methods (e.g., surveys, interviews ...

  20. Random Allocation & Random Selection

    Both random selection and random allocation are important parts of research and collecting data. Random selection is the method of selecting a sample from the population to participate in a study.

  21. Randomization in clinical studies

    9) Strictly speaking, the block size is randomly selected from a discrete uniform distribution, and so the use of a random block design rather than a "varying" block size would be a more formal procedure. 10) As the number of strata increases, the imbalance increases due to various factors. The details are beyond the scope of this study.

  22. PDF Selecting Research Participants

    psychological research, and this is not a huge problem, but random assignment of participants to groups is a very common procedure and is an important assumption of several statistical procedures. Random assignment means that participants have been independently assigned to groups. Imagine that we have selected 40 participants for a two-

  23. What's the difference between random assignment and random ...

    Random selection, or random sampling, is a way of selecting members of a population for your study's sample. In contrast, random assignment is a way of sorting the sample into control and experimental groups. Random sampling enhances the external validity or generalizability of your results, while random assignment improves the internal ...

  24. Module 3: Elements of Research

    Random selection is a form of sampling where a representative group of research participants is selected from a larger group by chance. This can be done by identifying all of the possible candidates for study participation (e.g., people attending the County fair on a Tuesday) and randomly choosing a subset to participate (e.g., selecting every ...

  25. Animals

    We analyzed 2836 randomly selected owner relinquishment records over a six-year period (2018-2023) using content analysis, a systematic method for categorizing qualitative data. The most common reasons identified for relinquishment included behavioral issues (28%), housing/moving (18%), and the inability to care for the animals (16%).

  26. Characterization of fresh biowastes for biogas production and

    Sampling of the fifteen representative study communities in Bayelsa State. Multistage random sampling was used to select fifteen representative communities for the study [].In the first sampling stage, six LGAs were randomly selected out of the eight LGAs in the state, as shown in Table 1.In the second stage, two communities were randomly selected from each of the selected LGAs, except for ...