Total sample size (n)
Population total = t = N * x
where N is the number of observations in the population, and x is the sample mean.
Or, if we know the sample proportion, we can estimate the population total (t) as:
Population total = t = N * p
where t is an estimate of the number of elements in the population that have a specified attribute, N is the number of observations in the population, and p is the sample proportion.
Sample mean = x = Σ( N h / N ) * x h
where N h is the number of observations in stratum h of the population, N is the number of observations in the population, and x h is the mean score from the sample in stratum h .
Sample proportion = p = Σ( N h / N ) * p h
where N h is the number of observations in stratum h of the population, N is the number of observations in the population, and p h is the sample proportion in stratum h .
Population total = t = ΣN h * x h
where N h is the number of observations in the population from stratum h , and x h is the sample mean from stratum h .
Or if we know the population proportion in each stratum, we can use this formula to estimate a population total:
Population total = t = ΣN h * p h
where t is an estimate of the number of observations in the population that have a specified attribute, N h is the number of observations from stratum h in the population, and p h is the sample proportion from stratum h .
x = ( N / ( n * M ) ] * Σ ( M h * x h )
where N is the number of clusters in the population, n is the number of clusters in the sample, M is the number of observations in the population, M h is the number of observations in cluster h , and x h is the mean score from the sample in cluster h .
p = ( N / ( n * M ) ] * Σ ( M h * p h )
where N is the number of clusters in the population, n is the number of clusters in the sample, M is the number of observations in the population, M h is the number of observations in cluster h , and p h is the proportion from the sample in cluster h .
Population total = t = N/n * ΣM h * x h
where N is the number of clusters in the population, n is the number of clusters in the sample, M h is the number of observations in the population from cluster h , and x h is the sample mean from cluster h .
And, if we know the sample proportion for each cluster, we can estimate a population total:
Population total = t = N/n * ΣM h * p h
where t is an estimate of the number of elements in the population that have a specified attribute, N is the number of clusters in the population, n is the number of clusters in the sample, M h is the number of observations from cluster h in the population, and p h is the sample proportion from cluster h .
s 2 = P * (1 - P)
where s 2 is an estimate of population variance, and P is the value of the proportion in the null hypothesis.
s 2 = Σ ( x i - x ) 2 / ( n - 1 )
where s 2 is a sample estimate of population variance, x is the sample mean, x i is the i th element from the sample, and n is the number of elements in the sample.
s 2 h = Σ ( x i h - x h ) 2 / ( n h - 1 )
where s 2 h is a sample estimate of population variance in stratum h , x i h is the value of the i th element from stratum h, x h is the sample mean from stratum h , and n h is the number of sample observations from stratum h .
s 2 h = Σ ( x i h - x h ) 2 / ( m h - 1 )
where s 2 h is a sample estimate of population variance in cluster h , x i h is the value of the i th element from cluster h, x h is the sample mean from cluster h , and m h is the number of observations sampled from cluster h .
s 2 b = Σ ( t h - t/N ) 2 / ( n - 1 )
where s 2 b is a sample estimate of the variance between sampled clusters, t h is the total from cluster h, t is the sample estimate of the population total, N is the number of clusters in the population, and n is the number of clusters in the sample.
You can estimate the population total (t) from the following formula:
where M h is the number of observations in the population from cluster h , and x h is the sample mean from cluster h .
SE = sqrt [ (1 - n/N) * s 2 / n ]
where n is the sample size, N is the population size, and s is a sample estimate of the population standard deviation.
SE = sqrt [ N 2 * (1 - n/N) * s 2 / n ]
where N is the population size, n is the sample size, and s 2 is a sample estimate of the population variance.
SE = (1 / N) * sqrt { Σ [ N 2 h * ( 1 - n h /N h ) * s 2 h / n h ] }
where n h is the number of sample observations from stratum h, N h is the number of elements from stratum h in the population, N is the number of elements in the population, and s 2 h is a sample estimate of the population variance in stratum h.
SE = sqrt { Σ [ N 2 h * ( 1 - n h /N h ) * s 2 h / n h ] }
where N h is the number of elements from stratum h in the population, n h is the number of sample observations from stratum h, and s 2 h is a sample estimate of the population variance in stratum h.
SE = | ( 1 / M ) * sqrt { [ N * ( 1 - n/N ) / n ] * Σ ( M * x - t / N ) / ( n - 1 ) |
+ ( N / n ) * Σ [ ( 1 - m / M ) * M * s / m ] } |
where M is the number of observations in the population, N is the number of clusters in the population, n is the number of clusters in the sample, M h is the number of elements from cluster h in the population, m h is the number of elements from cluster h in the sample, x h is the sample mean from cluster h, s 2 h is a sample estimate of the population variance in stratum h, and t is a sample estimate of the population total. For the equation above, use the following formula to estimate the population total.
t = N/n * Σ M h x h
With one-stage cluster sampling, the formula for the standard error reduces to:
SE = | ( 1 / M ) * sqrt { [ N * ( 1 - n/N ) / n ] * Σ ( M * x - t / N ) / ( n - 1 ) |
SE = | ( 1 / M ) * sqrt [ ( N * ( 1 - n/N ) / n ] * Σ ( M * p - t / N ) } / ( n - 1 ) |
+ ( N / n ) * Σ [ ( 1 - m / M ) * M * p * ( 1 - p ) / ( m - 1 ) ] } |
where M is the number of observations in the population, N is the number of clusters in the population, n is the number of clusters in the sample, M h is the number of elements from cluster h in the population, m h is the number of elements from cluster h in the sample, p h is the value of the proportion from cluster h, and t is a sample estimate of the population total. For the equation above, use the following formula to estimate the population total.
t = N/n * Σ M h p h
SE = | ( 1 / M ) * sqrt [ ( N * ( 1 - n/N ) / n ] * Σ ( M * p - t / N ) } / ( n - 1 ) |
SE = | N * sqrt { [ ( 1 - n/N ) / n ] * s /n + |
N/n * Σ ( 1 - m /M ) * M * s /m ) } |
where N is the number of clusters in the population, n is the number of clusters in the sample, s 2 b is a sample estimate of the variance between clusters, m h is the number of elements from cluster h in the sample, M h is the number of elements from cluster h in the population, and s 2 h is a sample estimate of the population variance in cluster h.
SE = N * sqrt { [ ( 1 - n/N ) / n ] * s 2 b /n }
When the null hypothesis is two-tailed, the critical value is the z-score or t-score that has a cumulative probability equal to 1 - α/2. When the null hypothesis is one-tailed, the critical value has a cumulative probability equal to 1 - α.
Researchers use a t-score when sample size is small; a z-score when it is large (at least 30). You can use the Normal Distribution Calculator to find the critical z-score, and the t Distribution Calculator to find the critical t-score.
If you use a t-score, you will have to find the degrees of freedom (df). With simple random samples, df is often equal to the sample size minus one.
Note: The critical value for a one-tailed hypothesis does not equal the critical value for a two-tailed hypothesis. The critical value for a one-tailed hypothesis is smaller.
UL = M + SE * CV
LL = M - SE * CV
The region of acceptance is the range of values between LL and UL. If the sample estimate of the population parameter falls outside the region of acceptance, the researcher rejects the null hypothesis. If the sample estimate falls within the region of acceptance, the researcher does not reject the null hypothesis.
By following the steps outlined above, you define the region of acceptance in such a way that the chance of making a Type I error is equal to the significance level .
In this section, two hypothesis testing examples illustrate how to define the region of acceptance. The first problem shows a two-tailed test with a mean score; and the second problem, a one-tailed test with a proportion.
As you probably noticed, defining the region of acceptance can be complex and time-consuming. Stat Trek's Sample Size Calculator can do the same job quickly, easily, and error-free.The calculator is easy to use, and it is free. You can find the Sample Size Calculator in Stat Trek's main menu under the Stat Tools tab. Or you can tap the button below.
An inventor has developed a new, energy-efficient lawn mower engine. He claims that the engine will run continuously for 5 hours (300 minutes) on a single ounce of regular gasoline. Suppose a random sample of 50 engines is tested. The engines run for an average of 295 minutes, with a standard deviation of 20 minutes.
Consider the null hypothesis that the mean run time is 300 minutes against the alternative hypothesis that the mean run time is not 300 minutes. Use a 0.05 level of significance. Find the region of acceptance. Based on the region of acceptance, would you reject the null hypothesis?
Solution: The analysis of survey data to test a hypothesis takes seven steps. We work through those steps below:
However, if we had to compute the sample mean from raw data, we could do it, using the following formula:
where Σx is the sum of all the sample observations, and n is the number of sample observations.
If we hadn't been given the standard deviation, we could have computed it from the raw sample data, using the following formula:
For this problem, we know that the sample size is 50, and the standard deviation is 20. The population size is not stated explicitly; but, in theory, the manufacturer could produce an infinite number of motors. Therefore, the population size is a very large number. For the purpose of the analysis, we'll assume that the population size is 100,000. Plugging those values into the formula, we find that the standard error is:
SE = sqrt [ (1 - 50/100,000) * 20 2 / 50 ]
SE = sqrt(0.9995 * 8) = 2.828
When the null hypothesis is two-tailed, the critical value has a cumulative probability equal to 1 - α/2. When the null hypothesis is one-tailed, the critical value has a cumulative probability equal to 1 - α.
For this problem, the null hypothesis and the alternative hypothesis can be expressed as:
Null hypothesis | Alternative hypothesis | Number of tails |
---|---|---|
μ = 300 | μ ≠ 300 | 2 |
Since this problem deals with a two-tailed hypothesis, the critical value will be the z-score that has a cumulative probability equal to 1 - α/2. Here, the significance level (α) is 0.05, so the critical value will be the z-score that has a cumulative probability equal to 0.975.
We use the Normal Distribution Calculator to find that the z-score with a cumulative probability of 0.975 is 1.96. Thus, the critical value is 1.96.
where M is the parameter value in the null hypothesis, SE is the standard error, and CV is the critical value. So, for this problem, we compute the lower limit of the region of acceptance as:
LL = 300 - 2.828 * 1.96
LL = 300 - 5.54
LL = 294.46
LL = 300 + 2.828 * 1.96
LL = 300 + 5.54
LL = 305.54
Thus, given a significance level of 0.05, the region of acceptance is range of values between 294.46 and 305.54. In the tests, the engines ran for an average of 295 minutes. That value is within the region of acceptance, so the inventor cannot reject the null hypothesis that the engines run for 300 minutes on an ounce of fuel.
Problem 2 Suppose the CEO of a large software company claims that at least 80 percent of the company's 1,000,000 customers are very satisfied. A survey of 100 randomly sampled customers finds that 73 percent are very satisfied. To test the CEO's hypothesis, find the region of acceptance. Assume a significance level of 0.05.
However, if we had to compute the sample proportion (p) from raw data, we could do it by using the following formula:
where s 2 is the population variance when the true population proportion is P, and P is the value of the proportion in the null hypothesis.
For the purpose of estimating population variance, we assume the null hypothesis is true. In this problem, the null hypothesis states that the true proportion of satisfied customers is 0.8. Therefore, to estimate population variance, we insert that value in the formula:
s 2 = 0.8 * (1 - 0.8)
s 2 = 0.8 * 0.2 = 0.16
For this problem, we know that the sample size is 100, the variance ( s 2 ) is 0.16, and the population size is 1,000,000. Plugging those values into the formula, we find that the standard error is:
SE = sqrt [ (1 - 100/1,000,000) * 0.16 / 100 ]
SE = sqrt(0.9999 * 0.0016) = 0.04
Null hypothesis | Alternative hypothesis | Number of tails |
---|---|---|
μ = 0.8 | μ < 0.8 | 1 |
Since this problem deals with a one-tailed hypothesis, the critical value will be the z-score that has a cumulative probability equal to 1 - α. Here, the significance level (α) is 0.05, so the critical value will be the z-score that has a cumulative probability equal to 0.95.
We use the Normal Distribution Calculator to find that the z-score with a cumulative probability of 0.95 is 1.645. Thus, the critical value is 1.645.
LL = 0.8 - 0.04 * 1.645
LL = 0.8 - 0.0658 = 0.7342
Thus, given a significance level of 0.05, the region of acceptance is the range of values between 0.7342 and 1.0. In the sample survey, the proportion of satisfied customers was 0.73. That value is outside the region of acceptance, so null hypothesis must be rejected.
Want to create or adapt books like this? Learn more about how Pressbooks supports open publishing practices.
Chapter 9: Survey Research
Learning Objectives
Survey research is a quantitative and qualitative method with two important characteristics. First, the variables of interest are measured using self-reports. In essence, survey researchers ask their participants (who are often called respondents in survey research) to report directly on their own thoughts, feelings, and behaviours. Second, considerable attention is paid to the issue of sampling. In particular, survey researchers have a strong preference for large random samples because they provide the most accurate estimates of what is true in the population. In fact, survey research may be the only approach in psychology in which random sampling is routinely used. Beyond these two characteristics, almost anything goes in survey research. Surveys can be long or short. They can be conducted in person, by telephone, through the mail, or over the Internet. They can be about voting intentions, consumer preferences, social attitudes, health, or anything else that it is possible to ask people about and receive meaningful answers. Although survey data are often analyzed using statistics, there are many questions that lend themselves to more qualitative analysis.
Most survey research is nonexperimental. It is used to describe single variables (e.g., the percentage of voters who prefer one presidential candidate or another, the prevalence of schizophrenia in the general population) and also to assess statistical relationships between variables (e.g., the relationship between income and health). But surveys can also be experimental. The study by Lerner and her colleagues is a good example. Their use of self-report measures and a large national sample identifies their work as survey research. But their manipulation of an independent variable (anger vs. fear) to assess its effect on a dependent variable (risk judgments) also identifies their work as experimental.
Survey research may have its roots in English and American “social surveys” conducted around the turn of the 20th century by researchers and reformers who wanted to document the extent of social problems such as poverty (Converse, 1987) [1] . By the 1930s, the US government was conducting surveys to document economic and social conditions in the country. The need to draw conclusions about the entire population helped spur advances in sampling procedures. At about the same time, several researchers who had already made a name for themselves in market research, studying consumer preferences for American businesses, turned their attention to election polling. A watershed event was the presidential election of 1936 between Alf Landon and Franklin Roosevelt. A magazine called Literary Digest conducted a survey by sending ballots (which were also subscription requests) to millions of Americans. Based on this “straw poll,” the editors predicted that Landon would win in a landslide. At the same time, the new pollsters were using scientific methods with much smaller samples to predict just the opposite—that Roosevelt would win in a landslide. In fact, one of them, George Gallup, publicly criticized the methods of Literary Digest before the election and all but guaranteed that his prediction would be correct. And of course it was. (We will consider the reasons that Gallup was right later in this chapter.) Interest in surveying around election times has led to several long-term projects, notably the Canadian Election Studies which has measured opinions of Canadian voters around federal elections since 1965. Anyone can access the data and read about the results of the experiments in these studies.
From market research and election polling, survey research made its way into several academic fields, including political science, sociology, and public health—where it continues to be one of the primary approaches to collecting new data. Beginning in the 1930s, psychologists made important advances in questionnaire design, including techniques that are still used today, such as the Likert scale. (See “What Is a Likert Scale?” in Section 9.2 “Constructing Survey Questionnaires” .) Survey research has a strong historical association with the social psychological study of attitudes, stereotypes, and prejudice. Early attitude researchers were also among the first psychologists to seek larger and more diverse samples than the convenience samples of university students that were routinely used in psychology (and still are).
Survey research continues to be important in psychology today. For example, survey data have been instrumental in estimating the prevalence of various mental disorders and identifying statistical relationships among those disorders and with various other factors. The National Comorbidity Survey is a large-scale mental health survey conducted in the United States . In just one part of this survey, nearly 10,000 adults were given a structured mental health interview in their homes in 2002 and 2003. Table 9.1 presents results on the lifetime prevalence of some anxiety, mood, and substance use disorders. (Lifetime prevalence is the percentage of the population that develops the problem sometime in their lifetime.) Obviously, this kind of information can be of great use both to basic researchers seeking to understand the causes and correlates of mental disorders as well as to clinicians and policymakers who need to understand exactly how common these disorders are.
Disorder | Average | Female | Male |
---|---|---|---|
Generalized anxiety disorder | 5.7 | 7.1 | 4.2 |
Obsessive-compulsive disorder | 2.3 | 3.1 | 1.6 |
Major depressive disorder | 16.9 | 20.2 | 13.2 |
Bipolar disorder | 4.4 | 4.5 | 4.3 |
Alcohol abuse | 13.2 | 7.5 | 19.6 |
Drug abuse | 8.0 | 4.8 | 11.6 |
And as the opening example makes clear, survey research can even be used to conduct experiments to test specific hypotheses about causal relationships between variables. Such studies, when conducted on large and diverse samples, can be a useful supplement to laboratory studies conducted on university students. Although this approach is not a typical use of survey research, it certainly illustrates the flexibility of this method.
Key Takeaways
Discussion: Think of a question that each of the following professionals might try to answer using survey research.
A quantitative approach in which variables are measured using self-reports from a sample of the population.
Participants of a survey.
Research Methods in Psychology - 2nd Canadian Edition Copyright © 2015 by Paul C. Price, Rajiv Jhangiani, & I-Chant A. Chiang is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License , except where otherwise noted.
Home » Survey Research – Types, Methods, Examples
Table of Contents
Definition:
Survey Research is a quantitative research method that involves collecting standardized data from a sample of individuals or groups through the use of structured questionnaires or interviews. The data collected is then analyzed statistically to identify patterns and relationships between variables, and to draw conclusions about the population being studied.
Survey research can be used to answer a variety of questions, including:
Survey Research Methods are as follows:
There are several types of survey research that can be used to collect data from a sample of individuals or groups. following are Types of Survey Research:
Based on Methodology Survey are divided into two Types:
Qualitative survey research.
Quantitative survey research is a method of collecting numerical data from a sample of participants through the use of standardized surveys or questionnaires. The purpose of quantitative survey research is to gather empirical evidence that can be analyzed statistically to draw conclusions about a particular population or phenomenon.
In quantitative survey research, the questions are structured and pre-determined, often utilizing closed-ended questions, where participants are given a limited set of response options to choose from. This approach allows for efficient data collection and analysis, as well as the ability to generalize the findings to a larger population.
Quantitative survey research is often used in market research, social sciences, public health, and other fields where numerical data is needed to make informed decisions and recommendations.
Qualitative survey research is a method of collecting non-numerical data from a sample of participants through the use of open-ended questions or semi-structured interviews. The purpose of qualitative survey research is to gain a deeper understanding of the experiences, perceptions, and attitudes of participants towards a particular phenomenon or topic.
In qualitative survey research, the questions are open-ended, allowing participants to share their thoughts and experiences in their own words. This approach allows for a rich and nuanced understanding of the topic being studied, and can provide insights that are difficult to capture through quantitative methods alone.
Qualitative survey research is often used in social sciences, education, psychology, and other fields where a deeper understanding of human experiences and perceptions is needed to inform policy, practice, or theory.
There are several Survey Research Data Analysis Methods that researchers may use, including:
Here are some common applications of survey research:
Here are some real-time examples of survey research:
Purpose of survey research.
The purpose of survey research is to gather data and insights from a representative sample of individuals. Survey research allows researchers to collect data quickly and efficiently from a large number of people, making it a valuable tool for understanding attitudes, behaviors, and preferences.
Here are some common purposes of survey research:
there are certain circumstances where survey research is particularly appropriate. Here are some situations where survey research may be useful:
Conducting survey research involves several steps that need to be carefully planned and executed. Here is a general overview of the process:
There are several advantages to using survey research, including:
Here are some of the main limitations of survey research:
Following is an example of a Survey Sample:
Welcome to our Survey Research Page! We value your opinions and appreciate your participation in this survey. Please answer the questions below as honestly and thoroughly as possible.
1. What is your age?
2. What is your highest level of education completed?
3. What is your current employment status?
4. How often do you use the internet per day?
5. How often do you engage in social media per day?
6. Have you ever participated in a survey research study before?
7. If you have participated in a survey research study before, how was your experience?
8. What are some of the topics that you would be interested in participating in a survey research study about?
……………………………………………………………………………………………………………………………………………………………………………………………………………………………………………………………………….
9. How often would you be willing to participate in survey research studies?
10. Any additional comments or suggestions?
Thank you for taking the time to complete this survey. Your feedback is important to us and will help us improve our survey research efforts.
Researcher, Academic Writer, Web developer
It might help to listen to people when asking them questions..
Posted June 8, 2024 | Reviewed by Gary Drevitch
One of the most commonly-used measurement tools in psychology research and practice is what’s called the patient-reported, or self-reported, test. This method involves giving a patient, client, or research study participant a survey or questionnaire to gauge their experience. The survey or questionnaire often asks about how someone feels or what they think. Sometimes, the questions can be about physical symptoms such as pain, fatigue, and poor sleep.
A trend I have noticed is that surveys and questionnaires are often denigrated with pejorative terms like “subjective” or “unreliable." The implications of these criticisms are that asking people to report their feelings, symptoms and thoughts is not worthwhile—or at least is less worthwhile than more “objective” tools like blood tests and brain scans. The definition of subjective is that it is something based on a person’s feelings while something objective is unbiased or based on facts.
This second-class status of self-reported information is highly detrimental to patients and the public in general for several reasons.
A primary reason why the denigration of surveys and questionnaires is harmful is that it is based on a myth. Despite claiming to be objective, the opinion that survey and questionnaire data is not useful is, in fact, not based on facts and is, instead, subjective .
Let’s consider some of the other types of tests that psychologists use in research or that physicians may use in clinical practice. The first is blood tests measuring chemicals in a person’s body such as stress hormones (cortisol), signs of inflammation (c-reactive protein), and metabolism (glucose). Other tests include body scans of the body ranging from X-rays to magnetic resonance imaging. Another common test is to ask the clinician how a patient is doing; for example, asking them to rate a patient’s function in a certain area on a seven-point scale. A common myth is that all these different tools (blood tests, scans, clinician report, surveys, and questionnaires) are measuring the same thing. They are not. What is in someone’s blood or shown on a scan is not the same thing as the emotions they are experiencing, the symptoms they have, or the thoughts running through their mind. Someone’s emotions, thoughts, and symptoms can differ greatly from what their physician or psychologist believes they are experiencing. The myth that all these tests measure the same thing forms the basis for disregarding self-reported data.
Another myth about surveys and questionnaires versus other types of tests is that other types are more reliable. This is not necessarily the case. There is a laundry list of blood tests that are unreliable but still used. Surveys and questionnaires, just like other types of data, have to go through a process of development to ensure that they are reliable—and they sometimes may be more reliable than other forms of data. What is measured by surveys and questionnaires—such as emotions, thoughts, symptoms)—can vary from hour to hour (or even minute to minute) and this natural variability is often mistaken for unreliability. This myth about surveys and questionnaires feeds into the previous myth that self-reported data is not worthwhile.
Despite the logical arguments for self-reported data, there is also a moral argument. Self-reported information from surveys and questionnaires is a way to measure what a person perceives about themselves. To disregard these perceptions in favor of other tests that might not measure the same thing is, to be blunt, paternalistic. It simply says that “I know better than you what you are experiencing based on this blood test/scan/my own opinion.” I’m not saying that these other tests should not be used; only that self-reported data should not be considered as less valuable than other tests.
In case you are wondering why I just spent an entire post waxing poetic about survey and questionnaire myths, there are several reasons. First, disregarding what people say on surveys and questionnaires is disregarding their experience and that is highly invalidating and harmful. Another problem is that time and money is often wasted chasing an “objective” measure of some psychological experience when we could have just asked people. Surveys and questionnaires are often much cheaper both to develop and use than blood tests, scans, and other biologically-based tests. While each type of test has its own place and value, self-reported surveys and questionnaires should definitely be considered as valuable as other measures.
Salene M. W. Jones, Ph.D., is a clinical psychologist in Washington State.
At any moment, someone’s aggravating behavior or our own bad luck can set us off on an emotional spiral that threatens to derail our entire day. Here’s how we can face our triggers with less reactivity so that we can get on with our lives.
A large volume of research highlights the adverse effects of relative deprivation on subjective well-being. Across different empirical settings and modelling approaches, a conceptual common denominator exists: the bulk of prior studies assumes that lower social status, by definition, implies higher relative deprivation, resulting in reduced well-being. In the present study, we take issue with this assumption and propose that lower self-ascribed positions on the status hierarchy are necessary but insufficient in and of themselves to undermine well-being. The critical, yet often neglected, factor in the literature is perceived societal unfairness. That is, one must believe that personal predicament as gauged by status disadvantage is, at least partly, due to some exogenous or impersonal forces (e.g., discrimination, limited opportunity). Our central argument is that the magnitude of the focal relationship between relative deprivation and well-being should be more pronounced among those who hold higher perceptions of unfairness. Using three independently collected probability datasets on the South Korean population—Social Science Korea (2017), Seoul Survey (2018), and Korean Social Integration Survey (2018)—we systematically test this hypothesis. Results from multilevel models robustly demonstrate that the connection between lower social status and lower well-being is significantly stronger among individuals who assess their society to be more ‘unfair,’ suggesting that future research should incorporate the level of perceived unfairness as a consequential moderator.
This is a preview of subscription content, log in via an institution to check access.
Price includes VAT (Russian Federation)
Instant access to the full article PDF.
Rent this article via DeepDyve
Institutional subscriptions
Data analyzed for this study are available upon reasonable request from the corresponding author.
Adjaye-Gbewonyo, K., & Kawachi, I. (2012). Use of the Yitzhaki Index as a test of relative deprivation for health outcomes: A review of recent literature. Social Science & Medicine, 75 , 129–137.
Article Google Scholar
Adler, N. E., Boyce, T., Chesney, M. A., Cohen, S., Folkman, S., Kahn, R. L., & Syme, S. L. (1994). Socioeconomic status and health: The challenge of the gradient. American Psychologist, 49 (1), 15–24.
Adler, N. E., Epel, E. S., Castellazzo, G., & Ickovics, J. R. (2000). Relationship of subjective and objective social status with psychological and physiological functioning: Preliminary data in healthy white women. Health Psychology, 23 , 586–592.
Alderson, A. S., & Katz-Gerro, T. (2017). Compared to whom? Inequality, social comparison, and happiness in the United States. Social Forces, 95 (1), 25–53.
Amendola, A., Dell’Anno, R., & Parisi, L. (2019). Happiness and inequality in European countries: Is it a matter of peer group comparisons? Economia Politica: Journal of Analytical and Institutional Economics, 36 (2), 473–508.
Anderson, C., Kraus, M. W., Galinsky, A. D., & Keltner, D. (2012). The local-ladder effect: Social status and subjective well-being. Psychological Science, 23 (7), 764–771.
Baldwin, M., & Mussweiler, T. (2018). The culture of social comparison. Proceedings of the National Academy of Sciences of the United States of America, 115 (39), E9067–E9074.
Google Scholar
Bárcena-Martín, E., Cortés-Aguilar, A., & Moro-Egido, A. I. (2017). Social comparison on subjective well-being: The role of social and cultural capital. Journal of Happiness Studies, 18 (4), 1121–1145.
Bjørnskov, C., Dreher, A., Fischer, J. A. V., Schnellenbach, J., & Gehring, K. (2013). Inequality and happiness: When perceived social mobility and economic reality do not match. Journal of Economic Behavior and Organization, 91 , 75–92.
Boyce, C. J., Brown, G. D. A., & Moore, S. C. (2010). Money and happiness: Rank of income, not income, affects life satisfaction. Psychological Science, 21 (4), 471–475.
Buttrick, N. R., Heintzelman, J. S., & Oishi, S. (2017). Inequality and well-being. Current Opinion in Psychology, 18 , 15–20.
Chan, C. H., Wong, H. K., & Yip, P. S. F. (2017). Associations of relative income deprivation with perceived happiness and self-rated health among the Hong Kong Chinese population. International Journal of Public Health, 62 (6), 697–707.
Chang, W.-C. (2013). Climbing up the social ladders: Identity, relative income, and subjective well-being. Social Indicators Research, 113 , 513–535.
Choi, M., & Park, H.-Y. (2022). The effect of household types on real estate asset inequality in Korea: Using the Survey of Household Finances and Living Condition Data. Journal of Urban Studies and Real Estate, 13 (3), 25–46. [In Korean].
Clark, A. E., Frijters, P., & Shields, M. A. (2008). Relative income, happiness, and utility: An explanation for the Eastrlin paradox and other puzzles. Journal of Economic Literature, 46 (1), 95–144.
Cojocaru, A. (2016). Does relative deprivation matter in developing countries: Evidence from six transition economies. Social Indicators Research, 125 , 735–756.
Delhey, J., & Dragolov, G. (2014). Why inequality makes Europeans less happy: The role of distrust, status anxiety, and perceived conflict. European Sociological Review, 30 , 151–165.
Delle Fave, A., Brdar, I., Wissing, M. P., Araujo, U., Castro Solano, A., Freire, T., Hernández-Pozo, M. D. R., Jose, P., Martos, T., Nafstad, H. E., Nakamura, J., Singh, K., & Soosai-Nathan, L. (2016). Lay definitions of happiness across nations: The primary of inner harmony and relational connectedness. Frontiers in Psychology, 7 , 30. https://doi.org/10.3389/fpsyg.2016.00030
Diener, E., Oishi, S., & Tay, L. (2018). Advances in subjective well-being research. Nature Human Behaviour, 2 , 253–260.
Diener, E., Pressman, S. D., Hunter, J., & Delgadillo-Chase, D. (2017). If, why, and when subjective well-being influences health, and future needed research. Applied Psychology: Health and Well-Being, 9 (2), 133–167.
Duesenberry, J. S. (1967). Income, saving, and the theory of consumer behavior . Harvard University Press.
Easterlin, R. A. 1974. Does economic growth improve the human lot? Some empirical evidence. In P. A. David, & W. R. Melvin (Eds.), Nations and households in economic growth. (pp. 89–125). New York, NY: Academic Press.
Eibner, C. E., & Evans, W. N. 2004. Chapter 14. The income-health relationship and the role of relative deprivation. In Neckerman, K. M (Ed.), Social Inequality (pp. 545–568). Russell Sage Foundation: New York.
Ferrer-i-Carbonell, A. (2005). Income and well-being: An empirical analysis of the comparison income effect. Journal of Public Economics, 89 , 997–1019.
Festinger, L. (1954). A theory of social comparison processes. Human Relations, 7 (2), 117–140.
Firebaugh, G., & Schroeder, M. B. (2009). Does your neighbor’s income affect your happiness? American Journal of Sociology, 115 (3), 805–831.
Gao, L., Sun, B., Du, Z., and Lv, G. 2022. How wealth inequality affects happiness: The perspective of social comparison. Frontiers in Psychology 13. doi: 10.3389/fpsyg.2022.829707.
García-Muñoz, T. M., Milgram-Baleix, J., & Odeh-Odeh, O. (2019). Inequality and life satisfaction in low- and middle-income countries: The role of opportunity. Societies, 9 (2), 37. https://doi.org/10.3390/soc9020037
Grosfeld, I., & Senik, C. (2010). The emerging aversion to inequality. Economics of Transition, 18 (1), 1–26.
Guven, C., & Sørensen, B. E. (2012). Subjective well-being: Keeping up with the perception of the Joneses. Social Indicators Research, 109 , 439–469.
Hofstede, G. (1980). Culture’s consequences: International differences in work-related values . Sage.
Hong, S., Kim, N., Mo, Z., & Yang, L. 2024. Income inequality in South Korea, 1933-2022: Evidence from distributional national accounts. World Inequality Lab Working Paper 2024/03.
Huang, Y. (2016). Downward Social Comparison Increases Life-Satisfaction in the Giving and Volunteering Context. Social Indicators Research, 125 , 665–676.
Hwang, S.-J. (2024). Subjective inequality in South Korea: Perception, belief, and discontent. Research in Social Stratification and Mobility, 89 , 100875.
Jeong, J. H., & Cheon, B. Y. (2017). Korea’s wealth inequality structure from an international perspective: Comparing with wealth inequalities in Korea, USA and Spain. The Geographical Journal of Korea, 51 (2), 149–164. [In Korean].
Kahneman, D., & Deaton, A. (2010). High income improves evaluation of life but not emotional well-being. Proceedings of the National Academy of Sciences of the United States of America, 107 (38), 16489–16493.
Katic, I., & Ingram, P. (2017). Income inequality and subjective well-being: Toward an understanding of the relationship and its mechanisms. Business & Society, 57 (6), 1010–1044.
Kelley, J., & Evans, M. D. R. (2017). The new income inequality and well-being paradigm: Inequality has no effect on happiness in rich nations and normal times, varied effects in extraordinary circumstances, increase happiness in poor Nations, and interacts with Individuals’ perceptions, attitudes, politics, and expectations for the future. Social Science Research, 62 , 39–74.
Kim, H.H.-S. (2022). Status perception, residential context and subjective wellbeing among South Korean adults: A multilevel analysis of primary survey data. The Social Science Journal, 59 (3), 405–423.
Kim, H. H-S. 2023. Changing perceptions of societal trust among Koreans: Relative deprivation, downward mobility, and sociotropic concern. In Kim, J (Ed.) A Contemporary Portrait of Life in Korea: Researching recent social and political trends (pp. 211–238) . Springer Nature Singapore.
Kitayama, S., Mesquita, B., & Karasawa, M. (2006). Cultural affordances and emotional experience: Socially engaging and disengaging emotions in Japan and the United States. Journal of Personality and Social Psychology, 91 (5), 890–903.
Lee, H., & Hong, D. (2023). The landscape of inequality in residential capitalism: Focusing on the marginalized impacts of income inequality on housing inequality. Housing Studies Review, 31 (4), 67–107. [In Korean].
Lee, M.-A., & Kawachi, I. (2017). The company you keep: Is socializing with higher status people bad for mental health? Sociology of Health & Illness, 39 (7), 1206–1226.
Lee, S., Shin, H., & Kim, C. (2020). Inequality of the household income and wealth in Korea: Research outcome and agenda. Economy and Society, 127 , 60–94. [In Korean].
Lee, W. (2018). Inequality of income and wealth in South Korea. Journal of Governmental Studies, 24 (2), 29–59. [In Korean].
Liao, T. F. (2021). Income inequality, social comparison, and happiness in the United States. Socius: Sociological Research for a Dynamic World, 7 , 1–17.
Lim, C. (2021). Chapter 11. Gong-jeong-seong-eun haeng-bog-eul ga-jyeo-da-jul su iss-na? (Can fairness bring happiness?) In Lee, H. & Oh, J. (Eds.) Gong-jeong-han sa-hoe-ui gil-eul mud-da (The Path to a Fair Society) . (pp. 392–417). Sigongsa: Seoul. [In Korean]
Markovits, D. (2019). The Meritocracy trap: How America’s foundational myth feeds inequality, dismantles the middle class, and devours the elite . Penguin Press.
Marmot, M., Smith, G. D., Stansfeld, S., Patel, C., North, F., Head, J., Ian, W., Brunner, E., & Feeney, A. (1991). Health inequalities among British civil servants: The Whitehall II Study”. Lancet, 337 , 1387–1393.
McBride, M. (2001). Relative-income effects on subjective well-being in the cross-section. Journal of Economic Behavior & Organization, 45 , 251–278.
Medvedev, O. N., & Landhuis, C. E. (2018). Exploring constructs of well-being, happiness and quality of life. PeerJ, 6 , e4903.
Mussweiler, T. (2003). Comparison processes in social judgment: Mechanisms and consequences. Psychological Review, 110 (3), 472–489.
Ng, Y. K. (2021). Happiness or life satisfaction? In Y. K. Ng (Ed.), Happiness—concept, measurement and promotion (pp. 33–39). Springer.
Olivos, F., Olivos-Jara, P., & Browne, M. (2021). Asymmetric social comparison and life satisfaction in social networks. Journal of Happiness Studies, 22 , 363–384.
Oshio, T., & Urakawa, K. (2014). The association between perceived income inequality and subjective well-being: Evidence from a social survey in Japan. Social Indicators Research, 116 , 755–770.
Park, K. (2021). Perception and characteristics of meritocracy in Korea. Journal of Citizen & World, 38 , 1–39. [In Korean].
Posel, D. R., & Casale, D. M. (2011). Relative standing and subjective wellbeing in South Africa: The role of perceptions, expectations and income mobility. Social Indicators Research, 104 (2), 195–223.
Quispe-Torreblanca, E. G., De Neve, J.-E., Brown, G. D. A., Boyce, C. J., & Wood, A. M. (2020). Inequality and social rank: Income increases buy more life satisfaction in more equal countries. Personality and Social Psychology Bulletin, 47 (4), 519–539.
Raila, H., Scholl, B. J., & Gruber, J. (2015). Seeing the world through rose-colored glasses: People who are happy and satisfied with life preferentially attend to positive stimuli. Emotion, 15 (4), 449–462.
Raudenbush, S. W., & Congdon, R. T. (2021). HLM 8: Hierarchical linear and nonlinear modeling . Scientific Software International Inc.
Ravazzini, L., & Chávez-Juárez, F. (2018). Which inequality makes people dissatisfied with their lives? Evidence of the link between life satisfaction and inequalities. Social Indicators Research, 137 (3), 1119–1143.
Reagan, P. B., Salsberry, P. J., & Olsen, R. J. (2007). Does the measure of economic disadvantage matter? Exploring the effect of individual and relative deprivation on intrauterine growth restriction. Social Science & Medicine, 64 (10), 2016–2029.
Sapolsky, R. M. (2005). The influence of social hierarchy on primate health. Science, 308 , 648–652.
Schalembier, B. (2018). An evaluation of common explanations for the impact of income inequality on life satisfaction. Journal of Happiness Studies: An Interdisciplinary Forum on Subjective Well-Being, 20 (3), 777–794.
Schneider, S. M. (2012). Income inequality and its consequences for life satisfaction: What role do social cognitions play? Social Indicators Research, 106 (3), 419–438.
Schneider, S. M. (2016). Income inequality and subjective wellbeing: Trends, challenges, and research directions. Journal of Happiness Studies, 17 , 1719–1739.
Schneider, S. M. (2019). Why income inequality is dissatisfying–perceptions of social status and the inequality-satisfaction link in Europe. European Sociological Review, 35 (3), 409–430.
Schwartz, S. H. (1999). A theory of cultural values and some implications for work. Applied Psychology: An International Review, 48 (1), 23–47.
Shavitt, S., Johnson, T. P., & Zhang, J. (2011). Horizontal and vertical cultural differences in the content of advertising appeals. Journal of International Consumer Marketing, 23 , 297–310.
Shifa, M., & Leibbrandt, M. (2018). Relative economic position and subjective well-being in a poor society: Does relative position indicator matter? Social Indicators Research, 139 , 611–630.
Smith, H. J., Pippin, G. M., Bialosiewicz, S., & Pettigrew, T. F. (2012). Relative deprivation: A theoretical and meta-analytic review. Personality and Social Psychology Review, 16 (3), 203–232.
StataCorp. (2013). Stata Statistical Software: Release 13 . StataCorp LP.
Subramanyam, M., Kawachi, I., Berkman, L., & Subramanian, S. V. (2009). Relative deprivation in income and self-rated health in the United States. Social Science & Medicine, 69 (3), 327–334.
Tan, J. J. X., Kraus, M. W., Carpenter, N. C., & Adler, N. E. (2020). The association between objective and subjective socioeconomic status and subjective well-being: A meta-analytic review. Psychological Bulletin, 146 (11), 970–1020.
Triandis, H. C., & Gelfand, M. J. (1998). Converging measurement of horizontal and vertical individualism and collectivism. Journal of Personality and Social Psychology, 74 (1), 118–128.
Ugur, Z. B. (2021). How does inequality hamper subjective well-being? The role of unfairness. Social Indicators Research, 158 , 377–407.
Wilkinson, R. G., & Pickett, K. E. (2006). Income inequality and population health: A review and explanation of the evidence. Social Science & Medicine, 62 (7), 1768–1784.
Wolff, L. S., Subramanian, S. V., Acevedo-Garcia, D., Weber, D., & Kawachi, I. (2010). Compared to whom? Subjective social status, self-rated health, and referent group sensitivity in a diverse US sample. Social Science & Medicine, 70 , 2019–2028.
Wood, J. V. (1989). Theory and research concerning social comparisons of personal attributes. Psychological Bulletin, 106 (2), 231–248.
Wood, J. V. (1996). What is social comparison and how should we study it? Personality and Social Psychology Bulletin, 22 (5), 520–537.
Wu, X., & Li, J. (2017). Income inequality, economic growth, and subjective well-being: Evidence from China. Research in Social Stratification and Mobility, 52 , 49–58.
Yang, X. Y., Hu, A., & Schieman, S. (2019). Relative deprivation in context: How contextual status homogeneity shapes the relationship between disadvantaged social status and health. Social Science Research, 81 , 157–169.
You, M., & Shim, H. (2013). A study on the cultural characteristics of Korean society: Discovering its categories using the cultural consensus model. Korean Journal of Culture and Social Issues, 19 (3), 457–485. [In Korean].
Download references
Authors and affiliations.
Department of Sociology, Ewha Womans University, Seoul, Republic of Korea
Ahhyun Cho & Harris Hyun-soo Kim
You can also search for this author in PubMed Google Scholar
Correspondence to Harris Hyun-soo Kim .
Conflicts of interest.
Authors declared no conflicts of interest with respect to the authorship or the publication of this article.
Publisher's note.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
Reprints and permissions
Cho, A., Kim, H.Hs. Perceived Unfairness Moderates the Association Between Relative Deprivation and Subjective Well-Being: Findings from an East Asian Country. Applied Research Quality Life (2024). https://doi.org/10.1007/s11482-024-10336-7
Download citation
Received : 16 January 2024
Accepted : 04 June 2024
Published : 19 June 2024
DOI : https://doi.org/10.1007/s11482-024-10336-7
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.
npj Mental Health Research volume 3 , Article number: 29 ( 2024 ) Cite this article
73 Accesses
22 Altmetric
Metrics details
Anxiety, a condition characterized by intense fear and persistent worry, affects millions each year and, when severe, is distressing and functionally impairing. Numerous machine learning frameworks have been developed and tested to predict features of anxiety and anxiety traits. This study extended these approaches by using a small set of interpretable judgment variables ( n = 15) and contextual variables (demographics, perceived loneliness, COVID-19 history) to (1) understand the relationships between these variables and (2) develop a framework to predict anxiety levels [derived from the State Trait Anxiety Inventory (STAI)]. This set of 15 judgment variables, including loss aversion and risk aversion, models biases in reward/aversion judgments extracted from an unsupervised, short (2–3 min) picture rating task (using the International Affective Picture System) that can be completed on a smartphone. The study cohort consisted of 3476 de-identified adult participants from across the United States who were recruited using an email survey database. Using a balanced Random Forest approach with these judgment and contextual variables, STAI-derived anxiety levels were predicted with up to 81% accuracy and 0.71 AUC ROC. Normalized Gini scores showed that the most important predictors (age, loneliness, household income, employment status) contributed a total of 29–31% of the cumulative relative importance and up to 61% was contributed by judgment variables. Mediation/moderation statistics revealed that the interactions between judgment and contextual variables appears to be important for accurately predicting anxiety levels. Median shifts in judgment variables described a behavioral profile for individuals with higher anxiety levels that was characterized by less resilience, more avoidance, and more indifference behavior. This study supports the hypothesis that distinct constellations of 15 interpretable judgment variables, along with contextual variables, could yield an efficient and highly scalable system for mental health assessment. These results contribute to our understanding of underlying psychological processes that are necessary to characterize what causes variance in anxiety conditions and its behaviors, which can impact treatment development and efficacy.
Introduction.
Anxiety disorders affected ~12% of the US population in 2021 1 and affects 4% of the population worldwide 2 , 3 . Anxiety disorders are characterized by intense fear and persistent worry in the absence of a defined threat 4 and are among the most common causes of disability worldwide. Anxiety disorders begin early in life 5 , 6 and increase the risk of subsequent mood disorders, substance misuse, suicidal behavior and economic disadvantage 7 .
The diagnosis of an anxiety disorder involves clinical determination of the severity of symptoms and the presence of specific symptom constellations based on clinical assessment, commonly augmented by surveys and symptom inventories 8 . Recently, automated approaches have been tested for predicting anxiety, as determined by clinical assessment or surveys, with a primary focus on using machine learning (ML) approaches 9 with large variable sets (e.g., >100) 10 , 11 , 12 including clinical data 13 , 14 , 15 , questionnaires 16 , 17 , wearable biosensors 18 , 19 , 20 , social media posts 21 , 22 , 23 , neural measures from MRI 24 , 25 , 26 and cognitive science variables 27 , 28 , 29 , 30 , 31 . These large variable sets add multiple dimensions to the characterization of anxiety across study participants producing higher accuracies and lower unexplained variance. They model complex relationships between the predictors and the outcome, yet can present challenges ranging from significant computational requirements and prohibitive privacy concerns, to lengthy and costly data acquisitions. The current study sought to contribute to current ML-based anxiety level prediction efforts by using a small set of cognitive science variables, that can be acquired in 2–3 min on a small digital device, like a smartphone. Currently, 92% of the US population 32 and 85% of the world population 33 can access such devices.
Cognitive science studies focused on judgment behavior are hypothesized to be relevant to anxiety given the overlap in the neural systems implicated in both 34 , 35 , 36 . Abnormalities in reward/aversion judgment have been linked to dopamine system dysfunction in depression, addiction, suicidality, and chronic stress 37 , 38 , 39 , and individuals with anxiety have shown salient alterations in reward/aversion judgment 40 , 41 , 42 . A number of reward/aversion variables are thought to represent biases in judgment 43 , 44 , such as loss aversion (LA) 45 and risk aversion (RA) 46 . Heightened RA 47 , 48 , 49 and heightened LA 50 , 51 have been reported in those with anxiety using a range of distinct monetary and emotional stimuli that describe reward/aversion judgment.
Reward/aversion judgment has been studied using operant keypress tasks to frame reinforcement reward in humans 52 , 53 , 54 , 55 , 56 , and been used to quantify judgment variables like LA 56 , 57 . These cognitive science studies have compared keypress-based LA to other LA frameworks such as prospect theory (e.g., Lee et al. 58 ), connected keypress methods to imaging of reward/aversion circuitry (e.g., refs. 52 , 59 , 60 , 61 ) and connected LA from operant keypressing to reward/aversion circuitry 62 . The keypress framework allows the modeling of human behavior using variance and entropic variables to produce a set of at least 15 features that characterize an individual’s reward/aversion judgment (e.g., LA, RA, and others, Table 1 ). These features have been linked to brain structure differences in the context of (1) substance use disorder 60 and (2) the characterization of substance use disorder and depression 63 .
These 15 judgment variables can also be computed from a picture rating task that takes 2–3 min 64 . The picture rating task was adapted from the operant keypress task, and is implementable on a smartphone or digital device (Fig. 1 , Table 1 ). Judgment variables from the shorter picture rating task are consistent across multiple data sets 64 , 65 . The 15 judgment variables derived from the picture rating task, when combined with a small set of demographic and survey variables, have been used to predict other mental health and medical health conditions with high accuracy using ML: depression history 65 , suicidality 66 , and vaccine uptake 67 . Based on results from these prior publications, we hypothesized that a small set of 15 judgment variables (see Fig. 1 , Table 1 ), with contextual variables theorized to affect judgment and mental function (in this case: demographics, perceived loneliness, and COVID-19 history), might facilitate the prediction of anxiety levels.
A An example picture from the picture rating task where participants were asked to rate how much they liked or disliked an imagine on a scale of −3 (dislike very much) and +3 (like very much), with 0 being neutral. B Visual representation of the x–y plane for relative preference theory (RPT) value function fitting and resulting features extracted. C Visual representation of the x–y plane for RPT limit function fitting and resulting features extracted. D Visual representation of the x–y plane for RPT tradeoff function fitting and resulting features extracted. E Each of the 15 features and their abbreviated terms.
A short picture rating task (see “Methods”) was administered to 4019 (3476 following data exclusion) de-identified participants in December 2021. Participants rated 48 unique color images from the International Affective Picture System (IAPS) 68 , 69 on a scale of −3 (dislike very much) to +3 (like very much). Anxiety scores were derived from the state component of the State-Trait Anxiety Inventory (STAI) questionnaire 70 , a validated anxiety questionnaire. Random Forest ( RF ) and balanced Random Forest ( bRF ) techniques were used to classify anxiety scores into ‘higher’ versus ‘lower’ classes and to understand the relative importance of the predictors using Gini scores. Post hoc mediation/moderation analyses were conducted to understand the interactions between judgment and contextual variables that may underly anxiety level prediction. Lastly, contextual and judgment variable differences were assessed against anxiety levels.
This study took the perspective that the power of psychological constructs depends on their capacity to make meaningful predictions. The use of mathematical cognitive science to predict survey-based anxiety measures contributes to our understanding of how psychological processes underlie the variance in anxiety conditions and behaviors, which might impact treatment development and efficacy.
This study assessed anxiety levels (derived from STAI questionnaire) in relation to contextual and 15 picture rating-derived judgment variables. Judgment variables include Loss Aversion (LA) 45 , Risk Aversion (RA) 46 , Loss Resilience (LR), Ante, Insurance, Total Reward Risk (Total RR), Total Aversion Risk (Total AR), Peak Positive Risk (Peak PR), Peak Negative Risk (Peak NR), Reward Tipping Point (Reward TP), Aversion Tipping Point (Aversion TP), Reward-Aversion tradeoff (RA tradeoff), Tradeoff range, Reward-Aversion consistency (RA consistency) and Consistency range (see Fig. 1B–D and Table 1 ).
For the classification of ‘higher’ and ‘lower’ anxiety levels, the bRF performed better than RF in terms of sensitivity, specificity, AUC ROC, and balanced accuracy at all three threshold values with the best performance for the threshold of 35 (Table 2 ). For the bRF classification, the out of bag (OOB) accuracy and accuracy ranged from 72 to 81% and AUC ROC ranged from 0.71 to 0.74 (Table 2 ) which was much higher than the chance levels obtained from the permutation analysis (Supplementary Table 3 ). The sensitivity ranged from 56 to 74% with the lowest values corresponding to a greater class imbalance at a threshold of 55. A greater class imbalance was noted as the threshold values increased as depicted in the ‘Percentage of data’ column of Table 2 . When the threshold value was 55, the percentage of participants with high anxiety was only 12% (418/3476) of the dataset.
For RF classification, the OOB accuracy and test dataset accuracy ranged from 72 to 88% and AUC ROC ranged from 0.52 to 0.72 (Table 2 ). The sensitivity ranged from 3 to 74% with the lowest values corresponding to a greater class imbalance at a threshold of 55. RF had the worst performance at the threshold of 55, with the output metrics close to the chance levels (Supplementary Table 3 ).
Multi-dimensional scaling (MDS) plots demonstrated how ‘higher’ and ‘lower’ clusters were better distinguished at lower STAI-S thresholds (i.e., Supplementary Fig. 2 ) and how bRF always produced better data segregation between the two classes as compared to RF .
Age, loneliness, income, and employment were consistently the most important individual features for bRF based on the mean decrease in Gini scores (see Fig. 2B, D, F ). This was also the case for RF with anxiety thresholds of 35 and 45 (Fig. 2A, C ). Together, these four variables contributed between 29 and 33% of the relative importance (Table 3 ). For the RF analysis with a high threshold of 55 (only 12% of the cohort in the ‘higher’ group and 3% sensitivity), age and loneliness remained the top-most contributing variables.
The predictors are arranged according to the mean decrease in Gini scores with the most important predictors on the top. The red box outlines the top contextual variables, and the blue box outlines the 15 judgment variables while the red * in the blue box points to the contextual variables in the cluster of judgment variables. Plots ( A ), ( C ), ( E ) corresponds to RF analyses with thresholds of 35, 45, and 55, respectively, and ( B ), ( D ), ( F ) corresponds to bRF with thresholds of 35, 45, and 55, respectively.
The 15 judgment variables contributed a combined 55–61% of the relative importance (Fig. 2 and Table 3 ). Other contextual variables (education group, education in years, marital status, race/ethnicity, sex, COVID-19 test, and diagnosis) were lower in classification importance, contributing a combined relative importance of 11–21%.
Mediation and moderation analysis were used to define statistical interactions between judgment variables and the most important contextual variables (age, loneliness, income, and employment). These contextual variables were defined as the mediator (Me) or moderator (Mo), judgment variables were defined as the independent variable, and STAI-S scores were defined as the dependent variable. These analyses revealed nine mediation (Table 4A and Supplementary Table 4 ) and seven moderation results (Table 4B ). Age acted as a mediator when loss resilience, Aversion TP, Tradeoff Range, and Consistency range were independent variables. Loneliness appeared as the mediator in four mediation analyses with ante, insurance, Total AR, and Consistency range, whereas employment only mediated Consistency range as the independent variable.
Independent and Me variables were then switched to test if the judgment variables acted as mediators. No significant mediation results were found when contextual variables were independent variables and judgment variables were mediators.
With regard to moderation analyses, age was found to also be involved in four moderations, with ante, insurance, Peak PR, and Peak NR as the independent variables. Loneliness was involved in one moderation with Peak PR, and employment was implicated in two moderations with insurance and Peak NR. There were no mediation or moderation results with the income variable.
Note that age, loneliness, and employment interacted with different judgment variables when acting as a mediators versus when acting as a moderators (see Table 4A and Table 4B ).
The seven demographic variables (excluding years of education), perceived loneliness, and COVID-19 history were assessed for differences across anxiety scores using Wilcoxon rank-sum and Kruskal Wallis tests. All contextual variables significantly varied by anxiety score ( p -value < 0.05) (Table 5A ). For the majority of contextual variables, boxplots depicted ascending or descending trends (Supplementary Fig. 3 ). For example, anxiety scores were higher (1) with higher levels of perceived loneliness, (2) among younger individuals, (3) in females, (4) among individuals with lower household income, (5) among individuals with lower education levels, and (6) in individuals reporting a history of COVID-19 infection (test and diagnosis).
Judgment variables were analyzed by ‘higher’ and ‘lower’ anxiety scores for the three threshold values (Fig. 3 ). Eleven out of the 15 judgment variables differed using the one-sided Wilcoxon rank sum test (significance α < 0.05) and 8 out 15 differed after correction for multiple comparisons (significance α < 0.0083, after Bonferroni correction). The alternative hypothesis, and the respective p -values for each test are reported in Table 5B . The alternative hypothesis was defined as the judgment variable distribution median being greater in the ‘higher’ anxiety group than the ‘lower’ anxiety group, or vice versa. The ‘higher’ anxiety group had higher medians for loss aversion (threshold = 35, p < 0.05), ante (threshold = 45, 55, p < 0.05), Peak PR (threshold = 45, p < 0.0083; 55, p < 0.05), and Total RR (threshold = 35, p < 0.05; 45, 55, p < 0.0083) when compared to the ‘lower’ anxiety group (Table 5B ). The ‘higher’ anxiety group had lower medians for risk aversion (threshold = 45, p < 0.05), loss resilience (threshold = 35, 45, 55, p < 0.0083), Peak NR (threshold = 35, p < 0.0083), Aversion TP (threshold = 35, 45, p < 0.0083), Total AR (threshold = 35, p < 0.0083), Tradeoff Range (threshold = 35, 45, 55, p < 0.0083), and RA consistency (threshold = 35, p < 0.05; 45, p < 0.0083) (Table 5B ). Insurance, Reward TP, RA Tradeoff, and Consistency Range showed no significant differences across all threshold values.
The thresholds 35, 45, and 55 roughly corresponds to 50th percentile (median), 75th percentile, and 90th percentile respectively of the anxiety/STAI-S scores. All values below the threshold are considered in ‘lower’ group and values above and equal to threshold are in ‘higher’ group.
This study evaluated how well a small set of judgment and contextual variables (i.e., demographics, perceived loneliness, and COVID-19 infection history), could together predict state anxiety levels. The study produced four major findings. First, prediction accuracy ranged from 72.4 to 81.3% with balanced Random Forest ( bRF) , and sensitivities decreased (74.1% to 56.1%) as the threshold for classifying anxiety was increased. Regardless of the threshold change, all prediction models maintained a relatively high AUC ROC (0.72 to 0.71), comparable with findings in the literature, and distinct from a permutation analysis showing AUC ROC outcomes approximating 0.50. bRF produced uniformly higher sensitivity outcomes than RF approaches. Second, four contextual variables (age, income, employment, and perceived loneliness) had the highest relative importance based on normalized Gini scores across most RF and bRF analyses (5 out of 6 analyses), and contributed a cumulative of 29–33% of relative importance to prediction. Other contextual variables such as race/ethnicity, sex, and COVID-19 infection history showed minimal importance across all analyses. All 15 judgment variables consistently showed similar importance and contributed a cumulative importance ranging from 55 to 61%. Third, nine of the 15 judgment variables were involved in mediation or moderation with contextual variables. Age, employment, and perceived loneliness mediated or moderated relationships with distinct judgment variables to model anxiety scores, consistent with other reports regarding the relationship of cognitive science measures and contextual variables 65 , 66 , 67 , 71 , 72 , 73 , 74 , 75 . Fourth, all contextual variables exhibited significant differences in anxiety scores, and 11 of the 15 judgment variables differed when assessed for median shifts across ‘higher’ and ‘lower’ anxiety groups, indicating that a constellation of judgment alterations are predictive of anxiety levels.
Prediction results from this study were comparable to recent research deploying advanced machine learning algorithms to predict anxiety and other mental health conditions 9 , with a number of limitations and advantages. Over the past decade there have been six general types of data used for predicting anxiety, with considerable heterogeneity in terms of how anxiety is defined, and the machine learning algorithms used. These six general frameworks for prediction included variables from: neuroimaging, physiological signals, survey-based assessments, social media posts, clinical or medical health records and behavioral tasks. Neuroimaging studies targeting anxiety prediction reported high accuracies ranging between 79 and 90% 24 , 25 and low correlation r = 0.28 (using Gaussian Process Regression) 26 but present challenges surrounding the collection of expensive, computationally-intensive, and complex MRI data that requires supervision of trained individuals at specific imaging sites. With similar caveats, studies using bio-signals and physiological signals 18 , 19 , 20 , that involved the use of multiple wearable sensors and collection of data under supervision, reported higher accuracies and model fits of 84.3%, 89.8% and r = 0.81. Survey-based studies have utilized extensive sets of demographic variables and lengthy questionnaires 16 , 17 , to predict anxiety with sensitivities between 62 and 73%. Other studies using demographics, lifestyle, and health surveys 10 , 11 , and transcripts of recorded interviews 12 predicted self-reported anxiety with accuracies between 75–86%. Studies using social media platforms like Reddit predicted anxiety with 75% precision 21 , 78% accuracy 22 ) from posts in mental health discussion groups, and Gruda et al. 23 used tweets to predict anxiety scores based on the level of anxiety assessed by volunteers in those tweets. Some studies 13 , 14 , 15 required access to clinical and medical records of thousands of participants and reported accuracies of 73–89% 11 , 13 , 14 , 18 , 20 , 25 .
A number of studies have utilized cognitive science variables derived from behavioral tasks to study anxiety 27 , 28 , 31 and neuroticism 29 (a general trait that is considered as a vulnerability factor for anxiety 76 ). For instance, Yamamori et al. 27 used approach-avoidance reinforcement learning tasks with hierarchical logistic regression (reporting p < 0.05) to model task-induced anxiety. Aupperle et al. 28 reported significant correlations ( p < 0.01) between measures from computer-based approach-avoidance conflict task and self-reported anxiety measures from anxiety sensitivity index and Behavioral Inhibition/Activation Scale. Park et al. 29 reported attenuated processing of gains and losses (using functional MRI responses to a Monetary Incentive Delay task) with higher polygenic risk scores for neuroticism using a general linear model ( p < 0.001). Forthman et al. 30 predicted repetitive negative thinking (a trait that negatively impacts anxiety) from 20 principal components of behavioral and cognitive variables (derived from detailed neuropsychological and behavioral assessment) and polygenic risk scores using a machine learning ensemble method with R 2 of 0.037 (standard error = 0.002). Richter et al. 31 utilized thorough behavioral testing completed by participants under supervision, and reported a sensitivity of 71.4% and a specificity of 70.8% using RF in individuals with anxiety and/or depression. The current study complements these publications, and supports their findings by using a machine learning-based approach and a short cognitive science task that can be performed without supervision on personal electronic device to predict anxiety with high accuracy and sensitivity.
Four contextual variables (age, loneliness, income, and employment status) were salient for the prediction of anxiety, having a cumulative relative importance of 29–33%. Although the relationship between anxiety levels and demographic measures was consistent with the literature (as described below), most of the other demographic measures did not contribute as much to anxiety level prediction. For instance, sex (gender assigned at birth) consistently contributed less than 1% of relative importance and race/ethnic background contributed 1–1.5% of relative importance across analyses. The 15 judgment variables contributed a cumulative relative importance ranging from 55 to 61%. These variables quantify irrationality, or biases, in judgment (i.e., the bounds to rationality as described by Kahneman 44 ), and support prior publications pointing to the importance of reward/aversion variables for the study anxiety 27 , 28 , 41 , 77 . Gini scores were minimally different across the 15 judgment variables, suggesting further research is needed to assess how these judgment variables interact or cluster together. When the threshold used for segregating ‘higher’ versus ‘lower’ anxiety groups was increased, education and marital status increased in feature importance for prediction, raising a hypothesis that some demographic variables may be more important for predicting severe anxiety. A history of COVID-19 infection was not salient for predicting current anxiety and was consistently one of the least important features. This result contrasts with other literature showing large-scale societal concern about COVID-19 illness, the pandemic and related anxiety 78 , 79 . This study only used the “state” and not the “trait” component from the STAI, to reflect current experience. The study thus does not address any relationship between prior COVID-19 history and long-term trait anxiety, nor address people’s thoughts about infection.
Mediation and moderation models were used to quantify relationships between contextual and judgment variables involved in the prediction of anxiety levels. Three of the four most important contextual variables (age, employment, and loneliness) interacted with judgment variables to predict, or model, anxiety scores in mediation and moderation frameworks. These relationships were not observed with income and no significant mediation were found when contextual variables (age, employment, income, and loneliness) acted as independent variables and judgment variables were the mediator variables. On the other hand, significant mediations with contextual variables as mediators indicated that the contextual variables statistically modulated the relationship between judgment variables and anxiety scores; that is, they sat in the causal pathway between judgment variables and anxiety levels. Seven unique judgment variables were involved in nine significant mediation models with age, employment, and loneliness as mediators. Moderation analyses reflected how an interaction between a contextual variable and judgment variable might predict anxiety scores; these relationships were observed for only four judgment variables across seven significant moderation models. In total, 9 of the 15 judgment variables were thus involved in either mediation or moderation, indicating that contextual variables affect the impact of a majority of the judgment variables on anxiety level prediction. From a psychological perspective, these findings demonstrate how context (e.g., age, employment status, perceived loneliness) modulates or interacts with judgment variables to model anxiety, and how these relationships between judgment and context may aide the assessment of anxiety and ultimately, other mental health conditions. Others have noted that psychological processes occur in a context, and this study supports their work 65 , 66 , 67 , 71 , 72 , 73 , 74 , 75 .
In the current study, anxiety scores significantly varied by the contextual variables used to classify anxiety scores. Anxiety scores increased with increasing levels of perceived loneliness, where participants who often, or always, avoid spending time with others, or spend most of their time alone, had higher levels of anxiety. This is consistent with previous literature where anxiety increased as a function of loneliness 80 , 81 , 82 and higher anxiety was related to avoidant social behavior 83 . Consistent with the literature, anxiety scores were predominately higher in females, as compared to males 84 , 85 , 86 , 87 , and in young adults (aged 18–39 years), as compared to older adults aged (40–70) 87 . As others have published, anxiety scores were higher in participants indicating lower household income levels 88 , 89 and lower education levels 90 . In alignment with other reports, anxiety scores also varied with different levels of employment where retired participants reported the lowest anxiety scores and participants that were unemployed 89 , 91 or had more than one job had the highest levels of anxiety. Anxiety scores also varied with marital status, in alignment with other reports, where participants classifying as ‘single’, ‘separated’, and ‘living with partner’ reported higher anxiety than others (e.g., ‘married’, ‘divorced’, ‘widowed’) 89 . As reported elsewhere our participants reporting mixed race backgrounds had higher anxiety 92 than other racial/ethnic groups (e.g., white, African American, Hispanic, Asian). Lastly, individuals who reported previous COVID-19 experienced more anxiety. This finding is consistent with other studies in adults 93 , 94 , 95 and a longitudinal study of adolescents with anxiety disorders that found SARS-COV-2 infection was associated with a 30% worsening in anxiety severity 96 , regardless of treatment status. Altogether, this concordance with the literature supports the broader set of findings.
Most judgment variables showed significant differences between ‘higher’ and ‘lower’ anxiety groups, suggesting three general constructs. The first being that the ‘higher’ anxiety group had higher loss aversion 50 , 51 which corresponds to an overweighting of bad outcomes relative to good ones 44 . The ‘higher’ anxiety group also had higher Peak Positive Risk and Total Reward Risk, indicating that there was a higher uncertainty that must be overcome to approach stimuli, and that the interactions between reward and the associated risk were higher. Both observations point to difficulties with initiating behavior toward positive things, per Markowitz’s decision utility equation 113 . The same participants had lower Peak Negative Risk as compared to the low anxiety group, indicating there was a lower uncertainty to avoid events. The ‘higher’ anxiety group also had lower Total Aversion Risk suggesting that interactions between aversion and the associated risk were lower. Similarly, they had lower Aversion Tipping Point which corresponds to the intensity of aversion beyond which avoidant choices are made, suggesting lower values in those with higher anxiety scores more readily make avoidant choices. Together, this set of judgment variables quantifies how individuals with high anxiety overweight bad outcomes relative to good ones, have difficulty approaching positive stimuli (i.e., more rewarding and non-aversive items), yet readily seem to avoid negative ones. A second construct suggests that individuals with difficulty approaching positive stimuli seem to be more open to risk-seeking. Participants with high anxiety scores had higher ante and lower risk aversion indicating they would be more willing to play a game with uncertain outcomes, and that they do not prefer actions that lead to certain outcomes (i.e., they prefer two birds in the bush vs. one in the hand). This contrasts with studies that used emotional and monetary stimuli to observe heightened risk aversion in individuals with anxiety 47 , 48 . This difference in observations might depend on how a question is placed in the context of gain or loss (i.e., framing effects) 44 , 97 . The third construct was identified by low loss resilience, tradeoff range, and consistency metrics. Specifically, a lower Tradeoff range in high-anxiety persons is consistent with a restrictive portfolio of positive and negative preferences. Lower Reward Aversion consistency suggests a likelihood of indifference or that a person neither likes nor dislikes a particular stimulus. People with high anxiety scores are less loss resilient, meaning they have a reduced ability to rebound from bad outcomes. The three constructs describe a behavioral profile for high anxiety persons as having less resilience, more avoidance, and more indifference behavior. Together, these three general groupings of judgment variables point to known features of anxiety 49 but provide a lawful, quantitative framework 56 , 57 for framing the condition and support the hypothesis that unique constellations of judgment variables underlie other mental health conditions like depression 65 and suicidality 66 . The current findings support calls for the development of a standard model of mind 98 , albeit based on processes of judgment and agency as opposed to variables focused primarily on cognition.
Several limitations need be considered. First, the participants were recruited from the United States, and region and culture may influence the importance of judgment variables in predicting anxiety as psychiatric symptoms differ across cultures 99 , 100 , 101 , 102 , 103 , 104 . Second, participants with mental health conditions were oversampled to meet criteria for other survey components not discussed here. This oversampling could bias results and more generalized samples are needed to validate and extend our findings. Third, all variables were self-reported and not collected from clinical records or framed as a double-blinded trial with investigator-administered survey instruments. Fourth, the cohort was sampled during the COVID-19 pandemic, in which greater incidents of loneliness and anxiety have been reported 78 , 79 . It will be important to prospectively investigate if similar behavioral patterns predict anxiety in the absence of a pandemic. Fifth, the survey did not request participants to differentiate between white non-Hispanic and non-white; more in-depth questions regarding racial and ethnic backgrounds should be considered in future data collections.
The current study used a computational cognition framework to assess how biases in human judgment might contribute to predicting anxiety levels. Using a small set of judgment and contextual variables (including demographics, perceived loneliness, and COVID-19 history) with a balanced Random Forest framework, this study achieved high accuracies up to 88.51% and AUC ROC values of 0.69–0.74 for predicting state anxiety levels derived from the STAI 70 . Judgment variables were extracted from a short (2–3 min), simple, and unsupervised picture rating task that can be easily completed on a personal electronic device. In these prediction analyses, the four most important variables (age, employment, income, and loneliness) were contextual variables that contributed 29–33% of the relative importance and judgment variables contributed up to 61% of the relative importance for prediction. Furthermore, age, loneliness, and employment status significantly mediated and moderated the relationship between judgment variables and anxiety scores—indicating statistically mechanistic relationships between these variables, and suggesting that both cognitive variables and contextual variables are important for accurately predicting anxiety levels. Judgment variables differed across participants with higher and lower anxiety scores providing a behavioral profile for participants with higher anxiety scores. That is to say, individuals with higher anxiety scores overweighted bad outcomes relative to good ones, had difficulty approaching positive stimuli, yet readily avoided negative ones. Along with this higher avoidance, they also had lower resilience and higher indifference, consistent with prior reports 49 . This study supports the hypothesis that a small set of interpretable judgment and contextual variables can accurately predict psychiatric symptoms and provide a computational cognitive framework to better understand and classify anxiety and other mental health conditions.
Gold Research Inc. (San Antonio, Texas) recruited study participants from multiple vendors in December 2021. 4019 de-identified participants (mean age ± std = 51.4 ± 14.9 years) were randomly sampled from the general U.S. population using an email survey database accessed by Gold Research, Inc. and a double opt-in methodology as described in detail in refs. 66 , 67 , 94 , 105 , 106 , and in the Supplemental Material. All participants provided informed consent following oversight by Northwestern University’s and the University of Cincinnati’s Institutional Review Board and in accordance with the Declaration of Helsinki (see “Ethical statement” and refs. 66 , 67 , 94 , 105 , 106 ). Participants were balanced to meet the U.S. Census Bureau’s demographic criteria at the time of the survey (December 2021) and oversampled by 15% of the sample for mental health conditions (see Supplemental Material). The survey was composed of several blocks of questions using questionnaires (detailed below) for depression, anxiety, suicidality, addiction, psychosis, violent ideation, disruptive and destructive behaviors, perceived loneliness, along with demographic, self-reported mental health, and COVID-19 history questionnaires. Participants also completed a 48-item picture rating task (Fig. 1 ) split into two 24 picture blocks.
Participation was offered with language noting that Gold Research was administering an emotional health questionnaire on behalf of Northwestern University, with the phrasing: “ We will be evaluating how different emotions and experiences are connected and may relate to our emotional health .” All participants provided written informed consent, including their primary participation in the study and the secondary use of their anonymized, de-identified (i.e., all identifying information removed by Gold Research Inc. prior to retrieval by the research group) data in secondary analyses (see Supplemental Material). The study was approved by the Institutional Review Boards for Northwestern University (NU) and University of Cincinnati (UC) in accordance with the Declaration of Helsinki (approval number STU00213665 for NU and 2023-0164 for UC).
Gold Research excluded participants using four criteria: (1) participants selected the same response throughout any section of the questionnaire (e.g., selecting option “1” for all questions), (2) participants indicated they had ten or more clinician-diagnosed illnesses out of a possible 17 (data not described here), (3) if both education level and years of education did not match, and (4) if they completed the questionnaire in less than 800 s. After filtering for these criteria, Gold Research provided the research team data from 4019 participants. These data were further screened using responses from the picture rating task. These procedures have been adapted from Azcona et al. 64 and are detailed in the Supplemental Material under Data filtering based on picture rating task . In short, participants were excluded if there was minimal variance in picture ratings (i.e., all pictures were rated the same or varied only by one point) and the quantitative feature set derived from the picture rating task was incomplete and/or there were extreme outliers (see Judgment variables from picture rating task and Supplemental Material). Using these exclusion criteria, 3476 participants were cleared for statistical analyses.
Participants completed the survey using the online platform provided by Gold Research, Inc. Participants were asked to self-report (a) perceived loneliness in the past month (loneliness), (b) demographics including age, gender assigned at birth (sex), annual household income (income), marital status (marital), employment status (employment), level of education (edu), number of years of education (edu_years), race/ethnicity (race/ethnicity), and (c) two COVID-19 questions: (i) if the participant had ever tested positive for COVID-19 (test) and (ii) if the participant was ever diagnosed by a clinician with COVID-19 (diagnosis). The complete text regarding these questions is listed under the Survey Questions section in Supplemental Material. The response set to (a)–(c) is referred to as ‘contextual variables’ hereafter. Following data filtering as described above and in refs. 94 , 105 , 106 , the 3476 participants were categorized as predominately female (61.5%), married (51.4%), white (85.7%), employed full-time (35.8%) with some college education (29.6%), and on average older (mean age = 51 years), see Supplementary Table 1 for a complete summary.
This study assessed anxiety in relation to contextual and picture rating-derived variables (described below). We used the State-Trait Anxiety Inventory (STAI) questionnaire which is commonly used to measure trait and state anxiety 70 . It is used in clinical settings to quantify anxiety. The STAI consists of 20 questions for current state anxiety, and 20 questions for trait anxiety. In this study, only the 20-state anxiety (STAI-S) questions were deployed in the online survey. Participants were instructed to answer each question based on a 4-point Likert scale (1 = Not at all; 2 = Somewhat; 3 = Moderately so; 4 = Very much so) based on how they feel right now , that is, at the time of the survey. The questions were scored following the instructions in the score key for the form Y-1 of STAI ( https://oml.eular.org/sysModules/obxOml/docs/ID_150/State-Trait-Anxiety-Inventory.pdf ). The scored sum of STAI-S ranged from 20 to 80 and is hereafter referred to as ‘STAI-S score’ and/or ‘anxiety score’. STAI-S score distributions are shown in Fig. 3 with three red arrows marking the threshold values used in classification as described under ‘Classification analysis’. The STAI thresholds of 35, 45, and 55 roughly corresponds to 50th percentile (median), 75th percentile and 90th percentile, respectively, of the STAI-S scores (Fig. 3 ).
Participants were shown 48 unique color images from the International Affective Picture System (IAPS) 68 , 69 . Six picture categories were used: (1) sports, (2) disasters, (3) cute animals, (4) aggressive animals, (5) nature (beach vs. mountains), and (6) adults in bathing suits, with eight pictures per category (48 pictures in total, a sample image is shown in Fig. 1A ), with all pictures in a category having similar published calibration. These images act as mildly emotional stimuli that are employed to assess both positive and negative value (i.e., reward or liking vs. aversion or disliking) and have been broadly used and validated in research of human emotion, attention, and preference 68 , 69 . Images were displayed on participants’ personal devices with a maximum size of 1204 × 768 pixels. Below each picture was a rating scale from −3 (dislike very much) to +3 (like very much), where 0 indicated a neutral point (Fig. 1A ). While there was no time limit for selecting a picture rating, participants were asked in the instructions to rate the images as quickly as possible and to use their first impression; specific instructions can be found in the Supplemental Material. Once a rating was selected, the next image was displayed.
Data from the picture rating task were analyzed using a computational framework to characterize preference judgments. Referred to as relative preference theory (RPT) 56 , 57 , 62 , this framework has been adapted to derive judgment features from picture ratings 64 , 66 , 67 as opposed to operant keypressing 52 , 56 , 57 , 59 , 60 , 61 , 62 , 107 . For each participant, picture ratings from each of the six image categories were split into two sets—positive and negative. For each of these two sets, and for all six categories, the mean, Shannon entropy 56 , 108 , and variance were calculated, yielding a tuple denoted as \(\left({{\boldsymbol{K}}}^{{\boldsymbol{+}}}{\boldsymbol{,}}{{\boldsymbol{\sigma }}}^{{\boldsymbol{+}}}{\boldsymbol{,}}{{\boldsymbol{H}}}^{{\boldsymbol{+}}}\right)\) for the positive ratings and \(\left({{\boldsymbol{K}}}^{{\boldsymbol{-}}}{\boldsymbol{,}}{{\boldsymbol{\sigma }}}^{{\boldsymbol{-}}}{\boldsymbol{,}}{{\boldsymbol{H}}}^{{\boldsymbol{-}}}\right)\) for the negative ratings. This resulted in a total of 36 \(\left({\boldsymbol{K}}{\boldsymbol{,}}{\boldsymbol{\sigma }}{\boldsymbol{,}}{\boldsymbol{H}}\right)\) variables. Next, for each participant, the mean across the six categories was computed for each \(\left({\boldsymbol{K}}{\boldsymbol{,}}{\boldsymbol{\sigma }}{\boldsymbol{,}}{\boldsymbol{H}}\right)\) variable, resulting in six additional variables: mean \(\left({{\boldsymbol{K}}}^{{\boldsymbol{+}}}{\boldsymbol{,}}{{\boldsymbol{\sigma }}}^{{\boldsymbol{+}}}{\boldsymbol{,}}{{\boldsymbol{H}}}^{{\boldsymbol{+}}}\right)\) representing reward behavior and mean \(\left({{\boldsymbol{K}}}^{{\boldsymbol{-}}}{\boldsymbol{,}}{{\boldsymbol{\sigma }}}^{{\boldsymbol{-}}}{\boldsymbol{,}}{{\boldsymbol{H}}}^{{\boldsymbol{-}}}\right)\) representing aversion behavior. For each participant, three separate curves for value, limit, and tradeoff functions (see Fig. 1B–D were plotted using MATLAB and the library polyfit , following other publications (see details in Supplemental Material) 56 , 57 , 62 , 64 , 107 . Representative curves from 500 randomly selected participants out of the 3476 cohort are shown in Supplementary Fig. 1 .
Goodness of fit for these functions was assessed by computing \({R}^{2}\) values, adjusted \({R}^{2}\) values (accounting for degrees of freedom), and \(F\) -statistics for each participant’s model fit (Supplementary Table 2A ). Individual participants’ \(\left({\boldsymbol{K}}{\boldsymbol{,}}{\boldsymbol{H}}\right)\) value functions were fit by concave logarithmic, or power-law functions (Supplementary Table 2 , Supplementary Fig. 1 ). \({R}^{2}\) values ranged from 0.85 to 0.94 for logarithmic fits of the value function, which was considered very high. Concave quadratic fits across individual participants’ \(\left({\boldsymbol{K}}{\boldsymbol{,}}{\boldsymbol{\sigma }}\right)\) data are displayed in Supplementary Fig. 1 , and goodness of fit assessed using the same metrics as with the \(\left({\boldsymbol{K}}{\boldsymbol{,}}{\boldsymbol{H}}\right)\) data (Supplementary Table 2A ). All \({R}^{2}\) values for the quadratic fits exceeded 0.80 and ranged from 0.84 to 0.96. Lastly, radial functions were fit to test for trade-offs in the distribution of \({{\boldsymbol{H}}}^{{\boldsymbol{-}}}\) and \({{\boldsymbol{H}}}^{{\boldsymbol{+}}}\) values across categories for each individual participant. Supplementary Fig. 1 displays radial fits across individual participants’ \(\left({{\boldsymbol{H}}}^{{\boldsymbol{+}}}{\boldsymbol{,}}{{\boldsymbol{H}}}^{{\boldsymbol{-}}}\right)\) and \(\left({{\boldsymbol{H}}}^{{\boldsymbol{+}}}{\boldsymbol{,}}{{\boldsymbol{H}}}^{{\boldsymbol{-}}}\right)\) data points for a random sample of participants.
The RPT framework fits reward/aversion curves and derives mathematical features from these graphical plots that are psychologically interpretable, scalable, recurrent, and discrete 56 , 57 , 64 . At least 15 variables can be extracted from this framework 64 , including Loss Aversion (LA) 45 , Risk Aversion (RA) 46 , Loss Resilience (LR), Ante, Insurance, Total Reward Risk (Total RR), Total Aversion Risk (Total AR), Peak Positive Risk (Peak PR), Peak Negative Risk (Peak NR), Reward Tipping Point (Reward TP), Aversion Tipping Point (Aversion TP), Reward-Aversion tradeoff (RA tradeoff), Tradeoff range, Reward-Aversion consistency (RA consistency) and Consistency range (Fig. 1B–D , Table 1 ). Each variable describes a quantitative component of the reward/aversion processing involved with judgment behavior (see details about each feature in Supplemental Material). The term ‘judgment variables’ will be used hereafter in reference to these features. Summary statistics for all 15 judgment variables obtained from all participants are summarized in Supplementary Table 2B .
All analyses were performed in R. Judgment variables and contextual variables, including demographics, perceived loneliness, and COVID-19 questions, were used in the classification analyses. Random Forest ( RF ) and balanced Random Forest ( bRF ) analyses were used to classify anxiety scores into ‘higher’ and ‘lower’ classes (see below). The open access package ‘randomForest’ in R was used to train the RF and bRF models on training dataset.
Anxiety scores were divided into two classes ‘higher’ and ‘lower’ classes based on three threshold values of 35, 45 and 55 as shown in Fig. 3 marked with red arrows. All values below a given threshold were labeled as ‘lower’ and values above and equal to the threshold were labeled as ‘higher’. Data were divided into train and test sets with a 70:30% ratio. RF and bRF approaches were implemented for each of the three thresholds using the command ‘randomForest’ from the package ‘randomForest’ in R. The number of variables randomly sampled as candidates at each split was 5 and the number of trees grown was 1000. The bRF performs random down-sampling of the majority class at each bootstrap sample to match the number of samples in majority and minority classes. The bRF approach was used in addition to the standard RF analysis because of the greater class imbalance when the anxiety score threshold was set to 45 and 55 (i.e., the ‘lower’ class occupied 70% and 88% of the dataset respectively). Note that bRF only performs down-sampling during training of the model. Once the model is trained, it calculates the prediction metrics on the complete, imbalanced train and test sets.
Out of bag (OOB) accuracy was reported for the training set. The model was then tested with the imbalanced test dataset and accuracy, sensitivity, specificity, AUC ROC (area under the receiving operating characteristics curve), and Balanced Accuracy (mean of sensitivity and specificity) were reported. Here, ‘higher’ was considered the positive class. For each threshold, percentages of each class relative to the entire dataset were reported. The entire procedure was repeated for each threshold value (i.e., 35, 45, 55). The multi-dimensional scaling (MDS) scaling coordinates of the proximity matrix for RF and bRF analyses were plotted to see how segregated the two clusters of lower and higher anxiety scores were, using the command ‘MDSplot’ in the package ‘randomForest’. The proximity matrix contains the frequency for each pair of data points. If two data points occupy the same terminal node through one decision tree, their proximity is increased by one. At the end of the run of all decision trees, the proximities are normalized by dividing by the number of trees. The MDS plots display how segregated the clusters were for the classification performed by RF or bRF .
To compare the performance of the classifiers to chance levels, permutation analysis was conducted for RF and bRF for each of the three threshold levels with 100 iterations. For each iteration, the ‘lower’/‘higher’ labels were shuffled randomly and RF and bRF analyses were run with the procedure described above. The above-mentioned output metrics were averaged over the 100 iterations to produce a model of chance effects.
The judgment and contextual variables were sorted based on the mean decrease in Gini scores and were plotted by decreasing feature importance, with the most important features appearing on the top of the plot. Gini score is a fundamental outcome of the random forest algorithms as it shows for each feature how large was the discriminative value for separating the data points into different classes. That is, how important was each variable in the classification, or how much was the uncertainty reduced in the model, leading to accurate predictions. The higher the mean decrease Gini score, the more important the feature was for the classification. The relative importance of the features was analyzed by normalizing the Gini score of each feature to the sum of all Gini scores. This gives the relative proportion of importance for each feature, with the sum for all features being 1. The Gini score plots, and the relative importance of the features were used in this study as an associated sensitivity analysis for the random forest algorithms.
Mediation and moderation were used as post hoc analyses to understand how judgment variables and the most important contextual variables, based on Gini scores, interact to model anxiety scores. These statistical mechanisms aide interpretation of the prediction results and follow procedures we have published before 109 , 110 .
Mediation was utilized to elucidate statistically causal relationship between judgment variables, contextual variables, and anxiety scores. The mediation model defines the relationship between an independent variable (X) and dependent variable (Y) with the inclusion of a third mediator variable (Me). Two sets of analyses were conducted where Y was the anxiety score and: (i) X were each of the judgment variables and Me were each of the most important contextual variables based on Gini score from RF and bRF analyses (i.e., the top four scores); (ii) X were the most important contextual variables and Me were the judgment variables.
The mediation model proposes that instead of a direct statistical causal relationship between X and Y, X influences Me, which then influences Y. Beta coefficients and their standard error (s) terms from the following linear regression equations, following the four-step process of Baron and Kenny (1986) 111 , 112 , were used to calculate Sobel p -values and mediation effect percentages ( T eff ):
Step 4 : Sobel’s test was then used to test if \({c}^{{\prime} }\) was significantly lower than \(c\) using the following equation:
Using a standard 2-tail z-score table, the Sobel p -value ( \({p}_{{Sobel}})\) was determined from the Sobel z-score and the mediation effect percentage ( T eff ) was calculated using the following equation:
Mediation was considered significant if p -values associated with terms a , b , and c were <0.05 from Step 1–3 and \({p}_{{Sobel}}\) < 0.05 111 and \({T}_{{eff}}\) > 50% 109 , 110 .
Secondary mediation analysis was run by switching variables assigned to X and Me to see if the mediation effects were directed. If \({p}_{{Sobel}}\) > 0.05 and \({T}_{{eff}}\) < 50% for the secondary mediation analysis, this supported that Me was in the causal pathway between X and Y.
The moderation model proposes that the strength and direction of the relationship between an independent variable (X) and dependent variable (Y) is controlled by the moderator variable (Mo). In this study, X were each of the judgment variables, Mo were each of the most important contextual variables based on Gini score from RF and bRF analysis, and Y was the anxiety score. Moderation is characterized by the interaction term between X and Mo in the linear regression equation as given below:
Moderation was considered significant if \({p}_{{\beta }_{3}}\le 0.05\) (the interaction term \({\beta }_{3}\) is significantly different than zero) and \({p}_{{overall}}\le 0.05\) (for the overall model) 109 , 110 . To check if the overall model was significant, we used F-test.
To test if the coefficient of the interaction term ( \({\beta }_{3})\) was significantly different than zero, we built full and restricted models and used partial F-tests to test the null hypothesis.
Full model:
Restricted Model:
Null hypothesis:
Alternative hypothesis:
If \({p}_{{\beta }_{3}}\) , associated with the partial F-test was less than 0.05, we rejected our null hypothesis regarding the interaction term.
These post hoc analyses assessed if the contextual variables and judgment variables differed by anxiety score.
Anxiety scores were assessed for differences by the different levels of contextual variables (except years of education) using Wilcoxon rank-sum test for questions with two levels and Kruskal Wallis test for questions with more than two levels. The ten contextual variables tested were loneliness, age, sex, income, marital status, employment status, education level, ethnicity, and COVID-19 test and diagnosis. Boxplots and p -values were reported.
Since judgment variables were continuous, they were divided into corresponding ‘higher’ and ‘lower’ groups following the respective grouping of anxiety scores at each of the three thresholds (35, 45, and 55), and tested using the one-sided Wilcoxon rank-sum test. Alternative hypotheses for each test, and the respective p -values, were reported. Bonferroni correction was done across all six tests (two tests for each of the three thresholds) for each judgment variable. The alternative hypothesis indicated if the judgment variable distributions differed between participants in ‘higher’ and ‘lower’ anxiety classes (for example, if a given judgment variable was higher in the ‘higher’ anxiety class as compared to the ‘lower’ anxiety class, or vice versa).
Data were de-identified before being provided to the investigators. Data are available in Microsoft Excel format and include relative preference variables, demographic metrics and survey variables inclusive of anxiety variables. The data may be accessed in Appendix 1 , Supplementary Information.
Computational behavior analysis used code published in refs. Azcona et al. 2022 and Kim et al. 2010. ML analyses used parameters as detailed in the Methods and Appendix 2 , Supplementary Information. Mediation/moderation analyses used code sequences as detailed in refs. Bari et al. 2021 and Vike et al. 2022.
CDC, FastStats: Mental Health. Center for Disease Control and Prevention (CDC) . https://www.cdc.gov/nchs/fastats/mental-health.htm (2021).
Ritchie, H. & Dattani, M. R. S. Mental Health - Our World in Data. Published online at OurWorldInData.org. https://ourworldindata.org/mental-health#risk-factors-for-mental-health (2021).
G. 2019 M. D. C.-T. L. Psychiatry, undefined 2022, Global, regional, and national burden of 12 mental disorders in 204 countries and territories, 1990–2019: a systematic analysis for the Global Burden of Disease Study 2019. Lancet Psychiatry 9 , 137–150 (2022).
Article Google Scholar
Chand, S. P. & Marwaha, R. Anxiety - StatPearls - NCBI Bookshelf (StatPearls, 2021).
Beesdo, K., Pine, D. S., Lieb, R. & Wittchen, H. U. Incidence and risk patterns of anxiety and depressive disorders and categorization of generalized anxiety disorder. Arch. Gen. Psychiatry 67 , 47–57 (2010).
Article PubMed Google Scholar
Beesdo-Baum, K. & Knappe, S. Developmental epidemiology of anxiety disorders. Child Adolesc. Psychiatr. Clin. N. Am. 21 , 457–478 (2012).
Asselmann, E., Wittchen, H. U., Lieb, R. & Beesdo-Baum, K. Sociodemographic, clinical, and functional long-term outcomes in adolescents and young adults with mental disorders. Acta Psychiatr. Scand. 137 , 6–17 (2018).
Article CAS PubMed Google Scholar
Bystritsky, A., Khalsa, S. S., Cameron, M. E. & Schiffman, J. Current diagnosis and treatment of anxiety disorders. P T 38 , 30–44 (2013).
PubMed PubMed Central Google Scholar
Iyortsuun, N. K., Kim, S. H., Jhon, M., Yang, H. J. & Pant, S. A review of machine learning and deep learning approaches on mental health diagnosis. Healthcare 11 , 285 (2023).
Article PubMed PubMed Central Google Scholar
van Eeden, W. A. et al. Predicting the 9-year course of mood and anxiety disorders with automated machine learning: a comparison between auto-sklearn, naïve Bayes classifier, and traditional logistic regression. Psychiatry Res. 299 , 113823 (2021).
Pavlova, I. et al. Determinants of anxiety levels among young males in a threat of experiencing military conflict-Applying a machine-learning algorithm in a psychosociological study. PLoS ONE 15 , e0239749 (2020).
Article CAS PubMed PubMed Central Google Scholar
Demiris, G. et al. Spoken words as biomarkers: Using machine learning to gain insight into communication as a predictor of anxiety. J. Am. Med. Inform. Assoc. 27 , 929–933 (2020).
Dabek, F. & Caban, J. J. A neural network based model for predicting psychological conditions. Lect. Notes Comput. Sci. 9250 , 252–261 (2015).
Sau, A. & Bhakta, I. Predicting anxiety and depression in elderly patients using machine learning technology. Health. Technol. Lett. 4 , 238–243 (2017).
Nemesure, M. D., Heinz, M. V., Huang, R. & Jacobson, N. C. Predictive modeling of depression and anxiety using electronic health records and a novel machine learning approach with artificial intelligence. Sci. Rep. https://doi.org/10.1038/s41598-021-81368-4 (2021).
Simjanoski, M. et al. Lifestyle predictors of depression and anxiety during COVID-19: a machine learning approach. Trends Psychiatry Psychother. 44 , 44–45 (2022).
Google Scholar
Bokma, W. A. et al. Predicting the naturalistic course in anxiety disorders using clinical and biological markers: a machine learning approach. Psychol. Med. 52 , 57–67 (2022).
Katsis, C. D., Katertsidis, N. S. & Fotiadis, D. I. An integrated system based on physiological signals for the assessment of affective states in patients with anxiety disorders. Biomed. Signal Process Control 6 , 261–268 (2011).
Perpetuini, D. et al. Prediction of state anxiety by machine learning applied to photoplethysmography data. PeerJ 9 , e10448 (2021).
Ihmig, F. R. et al. On-line anxiety level detection from biosignals: machine learning based on a randomized controlled trial with spider-fearful individuals. PLoS ONE 15 , e0231517 (2020).
Thorstad, R. & Wolff, P. Predicting future mental illness from social media: a big-data approach. Behav. Res. Methods 51 , 1586–1600 (2019).
Kim, J., Lee, J., Park, E. & Han, J. A deep learning model for detecting mental illness from user content on social media. Sci. Rep. 10 , 11846 (2020).
Gruda, D. & Hasan, S. Feeling anxious? Perceiving anxiety in tweets using machine learning. Comput. Hum. Behav. 98 , 245–255 (2019).
Ball, T. M., Stein, M. B., Ramsawh, H. J., Campbell-Sills, L. & Paulus, M. P. Single-subject anxiety treatment outcome prediction using functional neuroimaging. Neuropsychopharmacology 39 , 1254–1261 (2014).
Hilbert, K., Lueken, U., Muehlhan, M. & Beesdo-Baum, K. Separating generalized anxiety disorder from major depression using clinical, hormonal, and structural MRI data: A multimodal machine learning study. Brain Behav. 7 , e00633 (2017).
Portugal, L. C. L. et al. Predicting anxiety from wholebrain activity patterns to emotional faces in young adults: a machine learning approach. Neuroimage Clin. https://doi.org/10.1016/j.nicl.2019.101813 (2019).
Yamamori, Y., Robinson, O. J. & Roiser, J. P. Approach-avoidance reinforcement learning as a translational and computational model of anxiety-related avoidance. Elife 12 , RP87720 (2023).
Aupperle, R. L., Sullivan, S., Melrose, A. J., Paulus, M. P. & Stein, M. B. A reverse translational approach to quantify approach-avoidance conflict in humans. Behav. Brain Res. 225 , 455–463 (2011).
Park, H. et al. Polygenic risk for neuroticism moderates response to gains and losses in amygdala and caudate: evidence from a clinical cohort. J. Affect Disord. 293 , 124–132 (2021).
Forthman, K. L. et al. Transdiagnostic behavioral and genetic contributors to repetitive negative thinking: a machine learning approach. J. Psychiatr. Res. 162 , 207–213 (2023).
Richter, T., Fishbain, B., Markus, A., Richter-Levin, G. & Okon-Singer, H. Using machine learning-based analysis for behavioral differentiation between anxiety and depression. Sci. Rep. 10 , 1–12 (2020).
Martin, M. Computer and Internet Use in the United States: 2018 American Community Survey Reports ( www.census.gov/acs ) (2021).
Pews Research Center. (2019) Pew Research Center: Mobile Fact Sheet. https://www.pewinternet.org/factsheet/mobile/#targetText=The .
Hartley, C. A. & Phelps, E. A. Anxiety and decision-making. Biol. Psychiatry 72 , 113–118 (2012).
Hartley, C. A. & Phelps, E. A. Changing fear: the neurocircuitry of emotion regulation. Neuropsychopharmacology 35 , 136–146 (2010).
Rangel, A., Camerer, C., & Montague, P. R. A framework for studying the neurobiology of value-based decision making. Nat. Rev. Neurosci. 9 , 545–556 (2008).
Pedersen, M. L. et al. Computational phenotyping of brain-behavior dynamics underlying approach-avoidance conflict in major depressive disorder. PLoS Comput. Biol. 17 , e1008955 (2021).
Ironside, M. et al. Approach-avoidance conflict in major depressive disorder: congruent neural findings in humans and nonhuman primates. Biol. Psychiatry 87 , 399–408 (2020).
Bogdan, R. & Pizzagalli, D. A. Acute stress reduces reward responsiveness: implications for depression. Biol. Psychiatry 60 , 1147–1154 (2006).
Dillon, D. G. et al. Peril and pleasure: an rdoc-inspired examination of threat responses and reward processing in anxiety and depression. Depress Anxiety 31 , 233–249 (2014).
Breiter, H. C. & Gasic, G. P. A General Circuitry Processing Reward/Aversion Information and Its Implications for Neuropsychiatric Illness. Cognitive Neurosciences III, Third Edition 1043–1065 (2004).
Breiter, H. C., Gasic, G. P. & Makris, N. in Complex Systems Science in Biomedicine , 763–810 (2007).
Tversky, A. & Kahneman, D. Judgment under uncertainty: heuristics and biases. Science 185 , 1124–1131 (1974).
Kahneman, D. & Tversky, A. On the interpretation of intuitive probability: a reply to Jonathan Cohen. Cognition 7 , 409–411 (1979).
Chen, X., Voets, S., Jenkinson, N. & Galea, J. M. Dopamine-dependent loss aversion during effort-based decision-making. J. Neurosci. 40 , 661–670 (2020).
Wang, S., Krajbich, I., Adolphs, R. & Tsuchiya, N. The role of risk aversion in non-conscious decision making. Front Psychol. 3 , 50 (2012).
Charpentier, C. J., Aylward, J., Roiser, J. P. & Robinson, O. J. Enhanced risk aversion, but not loss aversion, in unmedicated pathological anxiety. Biol. Psychiatry 81 , 1014–1022 (2017).
Lorian, C. N. & Grisham, J. R. Clinical implications of risk aversion: an online study of risk-avoidance and treatment utilization in pathological anxiety. J. Anxiety Disord. 25 , 840–848 (2011).
Giorgetta, C. et al. Reduced risk-taking behavior as a trait feature of anxiety. Emotion 12 , 1373–1383 (2012).
Charpentier, C. J., De Martino, B., Sim, A. L., Sharot, T. & Roiser, J. P. Emotion-induced loss aversion and striatal-amygdala coupling in low-anxious individuals. Soc. Cogn. Affect Neurosci. 11 , 569–579 (2016).
Xu, P. et al. Amygdala–prefrontal connectivity modulates loss aversion bias in anxious individuals. Neuroimage 218 , 116957 (2020).
Aharon, I. et al. Beautiful faces have variable reward value: fMRI and behavioral evidence. Neuron 32 , 537–551 (2001).
Strauss, M. M. et al. fMRI of sensitization to angry faces. Neuroimage 26 , 389–413 (2005).
CAS PubMed Google Scholar
Elman, I. et al. Probing reward function in post-traumatic stress disorder with beautiful facial images. Psychiatry Res. 135 , 179–183 (2005).
Levy, B. et al. Gender differences in the motivational processing of facial beauty. Learn Motiv 39 , 136–145 (2008).
Kim, B. W. et al. Recurrent, robust and scalable patterns underlie human approach and avoidance. PLoS ONE 5 , e10613 (2010).
Livengood, S. L. et al. Keypress-based musical preference is both individual and lawful. Front Neurosci. 11 , 136 (2017).
Lee, S. et al. The commonality of loss aversion across procedures and stimuli. PLoS ONE 10 , e0135216 (2015).
Perlis, R. H. et al. Association of a polymorphism near CREB1 with differential aversion processing in the insula of healthy participants. Arch. Gen. Psychiatry 65 , 882–892 (2008).
Makris, N. et al. Decreased volume of the brain reward system in alcoholism. Biol. Psychiatry 64 , 192–202 (2008).
Gasic, G. P. et al. BDNF, relative preference, and reward circuitry responses to emotional communication. Am. J. Med. Genet., Part B: Neuropsychiatr. Genet. 150 , 762–781 (2009).
Viswanathan, V. et al. Age-related striatal BOLD changes without changes in behavioral loss aversion. Front Hum. Neurosci. 9 , 1–12 (2015).
Vike, N. L. et al. Characterizing major depressive disorder and substance use disorder using heatmaps and variable interactions: The utility of operant behavior and brain structure relationships. PLoS ONE 19 , e0299528(2024).
Azcona, E. A. et al. Discrete, recurrent, and scalable patterns in human judgement underlie affective picture ratings. https://doi.org/10.48550/arxiv.2203.06448 (2022).
Stefanopoulos, H. et al. In ICBHI 2022 Proceedings (figshare, 2022).
Lalvani, S. et al. Predicting suicidality with small sets of interpretable reward behavior and survey variables. Nat. Ment. Health 2024 , 1–14 (2024).
Vike, N. L. et al. Predicting COVID-19 vaccination uptake using a small and interpretable set of judgment and demographic variables: cross-sectional cognitive science study. JMIR Public Health Surveill. 10 , e47979 (2024).
Lang, P. & Bradley, M. International affective picture system (IAPS): technical manual and affective ratings. Psychology https://www.unifesp.br/dpsicobio/adap/instructions.pdf (1999).
Lang, P. J., Margaret, M. B. & Cuthbert, B. N. International affective picture system (IAPS): Affective Ratings of Pictures and Instruction Manual. Technical Report A-8. University of Florida, Gainesville, FL (2008).
Spielberger, C. D. Manual for the State-Trait Anxiety Inventory (Consulting Psychologists, 1983).
Schafer, K. M. The Status of Suicidality Prediction Research: A Meta-Analysis . https://diginole.lib.fsu.edu/islandora/object/fsu:723003/datastream/PDF/download (2019).
Schafer, K. M. et al. Suicidal ideation, suicide attempts, and suicide death among Veterans and service members: a comprehensive meta-analysis of risk factors. Mil. Psychol. 34 , 129–146 (2022).
Schafer, K. M., Wilson, E. & Joiner, T. Traumatic brain injury and suicidality among military veterans: the mediating role of social integration. J. Affect Disord. 338 , 414–421 (2023).
Resnik, P., Foreman, A., Kuchuk, M., Musacchio Schafer, K. & Pinkham, B. Naturally occurring language as a source of evidence in suicide prevention. Suicide Life Threat Behav. 51 , 88–96 (2021).
Shing, H. C., Resnik, P. & Oard, D. W. A prioritization model for suicidality risk assessment. In Proceedings of the Annual Meeting of the Association for Computational Linguistics , 8124–8137 (2020).
Clark, L. A., Watson, D. & Mineka, S. Temperament, personality, and the mood and anxiety disorders. J. Abnorm Psychol. 103 , 103–116 (1994).
Breiter, H. C., Aharon, I., Kahneman, D., Dale, A. & Shizgal, P. Functional imaging of neural responses to expectancy and experience of monetary gains and losses. Neuron 30 , 619–639 (2001).
Hawes, M. T., Szenczy, A. K., Klein, D. N., Hajcak, G. & Nelson, B. D. Increases in depression and anxiety symptoms in adolescents and young adults during the COVID-19 pandemic. Psychol. Med. 52 , 3222–3230 (2022).
Lakhan, R., Agrawal, A. & Sharma, M. Prevalence of depression, anxiety, and stress during COVID-19 pandemic. J. Neurosci. Rural Pr. 11 , 519–525 (2020).
Narchal, R. & Mcdavitt, S. A. Loneliness and anxiety sensitivity: understanding behavioural avoidance in the lonely. Acta Psychopathol. 3 , 58 (2017).
Cacioppo, J. T. et al. Lonely traits and concomitant physiological processes: the MacArthur social neuroscience studies. Int. J. Psychophysiol. 35 , 143–154 (2000).
Heinrich, L. M. & Gullone, E. The clinical significance of loneliness: a literature review. Clin. Psychol. Rev. 26 , 695–718 (2006).
Hayward, C. & Wilson, K. A. Anxiety sensitivity: a missing piece to the agoraphobia-without-panic puzzle. Behav. Modif. 31 , 162–173 (2016).
Lewinsohn, P. M., Lewinsohn, M., Gotlib, I. H., Seeley, J. R. & Allen, N. B. Gender differences in anxiety disorders and anxiety symptoms in adolescents. J. Abnorm Psychol. 107 , 109–117 (1998).
McLean, C. P. & Anderson, E. R. Brave men and timid women? A review of the gender differences in fear and anxiety. Clin. Psychol. Rev. 29 , 496–505 (2009).
Pigott, T. A. Gender differences in the epidemiology and treatment of anxiety disorders. J. Clin. Psychiatry 60 , 4–15 (1999).
PubMed Google Scholar
Nakazato, K. & Shimonaka, Y. The Japanese State-Trait Anxiety Inventory: age and sex differences. Percept. Mot. Skills 69 , 611–617 (1989).
Ridley, M., Rao, G., Schilbach, F. & Patel, V. Poverty, depression, and anxiety: causal evidence and mechanisms. Science (1979) 370 , eaay0214 (2020).
CAS Google Scholar
Watterson, R. A., Williams, J. V. A., Lavorato, D. H. & Patten, S. B. Descriptive epidemiology of generalized anxiety disorder in Canada. Can. J. Psychiatry 62 , 24–29 (2017).
Brenes, G. A. Age differences in the presentation of anxiety. Aging Ment. Health 10 , 298–302 (2006).
Hiswåls, A.-S., Walander, A., Soares, J. & Macassa, G. Employment status employment status, anxiety and depression in a municipal context. Res Health Sci. 2 , 12 (2017).
Fisher, S., Reynolds, J. L., Hsu, W. W., Barnes, J. & Tyler, K. Examining multiracial youth in context: ethnic identity development and mental health outcomes. J. Youth Adolesc. 43 , 1688–1699 (2014).
Mazza, M. G. et al. Anxiety and depression in COVID-19 survivors: role of inflammatory and clinical predictors. Brain Behav. Immun. 89 , 594–600 (2020).
Woodward, S. F. et al. Anxiety, post-COVID-19 syndrome-related depression, and suicidal thoughts and behaviors in COVID-19 Survivors: cross-sectional study. JMIR Form. Res. 6 , e36656 (2022).
Dorri, M. et al. Psychological problems and reduced health-related quality of life in the COVID-19 survivors. J. Affect Disord. Rep. 6 , 100248 (2021).
Strawn, J. R. et al. The impact of COVID-19 infection and characterization of long COVID in adolescents with anxiety disorders: a prospective longitudinal study. J. Am. Acad. Child Adolesc. Psychiatry , https://doi.org/10.1016/j.jaac.2022.12.027 (2023).
Tversky, A. & Kahneman, D. The framing of decisions and the psychology of choice. Science 211 , 453–458 (1981).
Laird, J. E., Lebiere, C. & Rosenbloom, P. S. A standard model of the mind: toward a common computational framework across artificial intelligence, cognitive science, neuroscience, and robotics. AI Mag. 38 , 13–26 (2017).
Bemme, D. & Kirmayer, L. J. Global Mental Health: interdisciplinary challenges for a field in motion. Transcult. Psychiatry 57 , 3–18 (2020).
Kirmayer, L. J. The politics of diversity: pluralism, multiculturalism and mental health. Transcult. Psychiatry 56 , 1119–1138 (2019).
Kpanake, L. Cultural concepts of the person and mental health in Africa. Transcult. Psychiatry 55 , 198–218 (2018).
Weine, S. M., Langenecker, S. & Arenliu, A. Global mental health and the National Institute of Mental Health Research Domain Criteria. Int. J. Soc. Psychiatry 64 , 436–442 (2018).
Bhugra, D. Sexual variation and mental health. Nord J. Psychiatry 72 , S27–S29 (2018).
Bredström, A. Culture and context in mental health diagnosing: scrutinizing the DSM-5 revision. J. Med. Human. 40 , 347–363 (2019).
Bari, S. et al. The prevalence of psychotic symptoms, violent ideation, and disruptive behavior in a population with SARS-CoV-2 infection: preliminary study. JMIR Form. Res 6 , e36444 (2022).
Vike, N. L. et al. The relationship between a history of high-risk and destructive behaviors and COVID-19 infection: preliminary study. JMIR Form. Res. 7 , e40821 (2023).
Viswanathan, V. et al. A quantitative relationship between signal detection in attention and approach/avoidance behavior. Front Psychol. 8 , 122 (2017).
Shannon, C. E. & Weaver, W. The Mathematical Theory of Communication Vol. 1, 117 (University of Illinois Press, 1949).
Bari, S. et al. Integrating multi-omics with neuroimaging and behavior: a preliminary model of dysfunction in football athletes. Neuroimage: Rep. 1 , 100032 (2021).
Vike, N. L. et al. A preliminary model of football-related neural stress that integrates metabolomics with transcriptomics and virtual reality. iScience 25 , 103483 (2022).
Baron, R. M. & Kenny, D. A. The moderator-mediator variable distinction in social psychological research. conceptual, strategic, and statistical considerations. J. Pers. Soc. Psychol. 51 , 1173–1182 (1986).
Kenny, D. A. Mediation. http://davidakenny.net/cm/mediate.htm (2021).
Markowitz, H. The utility of wealth. J. Political Econ. 60 , 151–158 (1952).
Download references
We thank Carol Ross, Angela Braggs-Brown, Thomas M. Talavage, Eric Nauman, and Marc Cahay at College of Engineering and Applied Science, University of Cincinnati who significantly impacted the transfer of research funding to University of Cincinnati (UC), allowing this work to be completed. Funding for this work was provided by Office of Naval Research award N00014-21-1-2216 [H.C.B.], Office of Naval Research award N00014-23-1-2396 [H.C.B. (contact), A.K.K.] and Jim Goetz donation to the University of Cincinnati, College of Engineering and Applied Science (H.C.B.).
These authors contributed equally: Byoung-Woo Kim, Nicole L. Vike, Shamal Lalvani, Leandros Stefanopoulos.
These authors jointly supervised this work: Aggelos K. Katsaggelos, Hans C. Breiter.
Department of Computer Science, University of Cincinnati, Cincinnati, OH, USA
Sumra Bari, Byoung-Woo Kim, Nicole L. Vike & Hans C. Breiter
Department of Electrical Engineering, Northwestern University, Evanston, IL, USA
Shamal Lalvani, Leandros Stefanopoulos & Aggelos K. Katsaggelos
Laboratory of Medical Informatics, Aristotle University of Thessaloniki, Thessaloniki, Greece
Leandros Stefanopoulos & Nicos Maglaveras
Integrated Marketing Communications, Medill School of Journalism, Northwestern University, Evanston, IL, USA
Martin Block
Department of Psychiatry and Behavioral Neuroscience, College of Medicine, University of Cincinnati, Cincinnati, OH, USA
Jeffrey Strawn
Department of Computer Science, Northwestern University, Evanston, IL, USA
Aggelos K. Katsaggelos
Department of Radiology, Northwestern University, Chicago, IL, USA
Department of Biomedical Engineering, University of Cincinnati, Cincinnati, OH, USA
Hans C. Breiter
Department of Psychiatry, Massachusetts General Hospital and Harvard School of Medicine, Boston, MA, USA
You can also search for this author in PubMed Google Scholar
Study concept/design: H.C.B., A.K.K. and S.B. Acquisition of original data: H.C.B., B.W.K., S.B., N.L.V., S.L., L.S., M.B. and A.K.K. Coding of statistical tools: S.B. and B.W.K. (with guidance from H.C.B. and A.K.K.). Analysis of data: S.B. and B.W.K. (with guidance from H.C.B. and A.K.K.). Interpretation of data: S.B., H.C.B. and A.K.K. (with input from B.W.K., N.L.V., S.L., L.S., B.W.K., M.B. and N.M.). Statistical assessment: S.B. and B.W.K. (with guidance from H.C.B. and A.K.K.). Authored original draft: S.B. and H.C.B. Generated figures: S.B., B.W.K. and H.C.B. Revision of manuscript for content: all authors. All authors approved the final version of the paper for submission.
Correspondence to Hans C. Breiter .
Competing interests.
S.B., B.W.K., N.V., S.L., L.S., M.B., A.K. and H.C.B. submitted a provisional patent “Integrating Cognitive Science with Machine Learning for High Accuracy Prediction of Anxiety Disorders”. The provisional application is led by University of Cincinnati (Office of Innovation) in conjunction with Northwestern University, Application # 63/648,898. The other authors declare no competing interests.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplemental material, supplemental appendix 1, supplemental appendix 2, rights and permissions.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .
Reprints and permissions
Cite this article.
Bari, S., Kim, BW., Vike, N.L. et al. A novel approach to anxiety level prediction using small sets of judgment and survey variables. npj Mental Health Res 3 , 29 (2024). https://doi.org/10.1038/s44184-024-00074-x
Download citation
Received : 11 June 2023
Accepted : 16 May 2024
Published : 18 June 2024
DOI : https://doi.org/10.1038/s44184-024-00074-x
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.
INFORMATION FOR
Health limitations, distance, work schedules, family obligations, and financial constraints are all barriers that prevent patients from participating in clinical trials. For those living with Long COVID, debilitating symptoms can make traveling to a study site impossible.
Now, Harlan Krumholz, MD , Harold H. Hines, Jr. Professor of Medicine (Cardiology), is pioneering a new approach that makes participation in clinical research accessible for those whose lives have been upended by the post-acute infection syndrome. At Yale School of Medicine, working closely with Sterling Professor Akiko Iwasaki, PhD , he is the principal investigator of the Yale Paxlovid for Long COVID (PAX LC) Trial . This phase 2 investigational new drug clinical trial evaluating the use of the antiviral for people with Long COVID with a decentralized nationwide design that brings the research into participants’ homes. Operating throughout the contiguous United States, it is the first fully decentralized phase 2 trial with this complexity and scale.
Krumholz believes that this trial demonstrates that a future of clinical research that allows patients to participate at their convenience is not only possible, but more efficient and even cost-effective than standard clinical trials.
For example, April 9 marked Cindy’s* fourth anniversary of living with Long COVID.
It started with severe fatigue that made her feel as if her limbs “weighed a million pounds,” followed by shortness of breath and gastrointestinal issues after she was infected with COVID-19 in 2020. Her symptoms never went away. A single mom in Texas with a 6-year-old son, she says the disease forced her to get creative: She made up a game where he rolled a ball around her while she lay still on their trampoline.
Over the years, her symptoms have ebbed and flowed. She suffered a sharp decline in January 2022 — the height of the Omicron wave — when she was hit with her second acute COVID infection. This triggered the start of a slow decline that eventually took a toll on her cognitive abilities, hindering her ability to perform her job as an aerospace engineer and eventually forcing her to take a leave of absence in the fall of 2023. This was also when she began experiencing another common Long COVID symptom known as post-exertional malaise (PEM), in which her symptoms significantly worsened after physical or mental activity, sometimes leaving her bedbound for days at a time.
“I was known for thinking outside of the box, being able to juggle things, look to the future, and solve problems. I knew how to put big, complicated puzzles together,” she says. “Suddenly, everything seemed so complicated. Living became so hard. I was just trying to survive day-by-day, if not hour-by-hour, managing all of my symptoms.”
Desperate for answers, Cindy stayed on top of ongoing clinical trials. Last year, she applied for a stem cell therapy trial in Houston but was eliminated as a potential participant during the final screening. On the one hand, she was devastated. “I was absolutely convinced that the therapy was going to be the answer for me,” she says.
But enrolling in the trial would have presented its own challenges. She lives 45 minutes away from Houston and would have needed help getting rides. And she wondered if she had the capacity to be able to participate in all that the trial required. “From an accessibility standpoint, most clinical trials require you to be there in person for exams, bloodwork, tests, or whatever else they need,” she says. “That’s just not feasible when you have a limiting condition.”
Then, she came across that the PAX LC Trial. Typically, participating in a clinical trial with a Connecticut-based university would have been impossible. But the YSM team was prepared to bring the study to patients no matter where they lived in the U.S. Participants received the drug or placebo in their mailboxes and filled out electronic diaries. They gave blood and saliva samples at home or at a nearby lab. “The Pax LC trial is a historic contribution to the evolution of a new way of doing trials,” Krumholz says. “I’m really proud of what we’ve accomplished, and I’m hoping it will be a spark for the future.”
Researchers currently have several hypotheses for the underlying causes of Long COVID, including persistent virus, lingering SARS-CoV-2 viral remnants, autoimmune dysfunction, reactivated latent viruses like Epstein-Barr, and tissue damage. These hypotheses are not mutually exclusive, and it is possible that individuals with Long COVID may experience a combination of these mechanisms. Krumholz, Iwasaki, and their team designed the Pax LC Trial to test the persistent virus hypothesis.
Cindy initiated the prescreening process by filling out a survey about pre-existing conditions and syncing her medical records to an app on her phone. Several months later, she received an email inviting her to move forward. The next steps of the prescreening involved going to a local medical laboratory to give blood samples, as well as adjusting some of her medications.
Once enrolled, Cindy received her course of either Paxlovid or a placebo by mail. Each night, she answered survey questions about her symptoms from her phone. She also gave blood and saliva samples five times. Three of these times, the Pax LC team sent a technician to her home. “To be able to have somebody come to your house — that just made life so much easier,” she says. For the other two sample collection dates, she went to a nearby laboratory.
Despite the fact that our trial is sort of built as a concierge service for each individual patient, it was much less expensive than what we had been doing. Harlan Krumholz, MD
Patients aren’t the only ones who can benefit from decentralized trials, says Krumholz. For over a decade, he has been an advocate for reforming clinical research, arguing that its current structure stymies innovation. His colleagues in the laboratory are discovering potential avenues to treating diseases at a “dizzying pace,” he says. However, because standard clinical trials struggle to recruit and keep patients, they have not been able to keep up with the pace of scientific discovery, which prevents new drugs and medical devices from reaching the market.
“We’re in the midst of a life sciences revolution,” says Krumholz. “The central chokepoint is evidence generation — cycle times are too slow, and trials are so expensive that our level of confidence has to be very high, or no one will make the investment.”
This is because for many, the ability to participate in clinical trials and conform to the schedule they demand is a luxury. Not only does this slow enrollment, but it also contributes to the stark lack of diversity in research. “The current way of doing things is often leaving us in a position where we are having to chase participants and cajole them to stay in the studies that were built for their benefit,” says Krumholz. “Imagine how off-putting clinical trials must be when people who have the most to gain are left in the position where they don’t want to continue.”
The solution, he proposes, is creating clinical trials, modeled after PAX LC, that treat participants as partners and accommodate their constraints. PAX LC not only conformed to participants’ schedules, but also strived to create a sense of community through running virtual town halls for people with ong COVID so they could ask questions and stay updated on results as researchers got them. “People don’t have any obligation to join trials, so we need to make it so they’re something they want to join,” he says.
Ezra,* a freelance cartoonist in Minnesota, began experiencing COVID-19 symptoms in September 2022 and never recovered. At the time, he suffered PEM so severe that exertion as small as “microwaving lunch” could put him in bed for days. He saw doctor after doctor at his nearby hospital, and when that failed, he decided to try traveling to the Mayo Clinic. “I feel very fortunate that I was able to get somewhere like Mayo a few times, but it is an hour and a half drive from Minneapolis,” he says. “It cost my partner gas money and lost time at work [to take me there].”
In his search for answers, he learned of the Pax LC Trial and decided to enroll. “The effort to bring the trial into people’s houses is valiant, especially for a population that is chronically ill,” he says. “For disabled people to become a part of clinical trials already requires so much of us—even those of us who have support systems that are able to take care of us. Adding an extra thing to our plates is often just not feasible for most people with chronic illness.”
Other Pax LC Trial participants echo Ezra’s sentiments. “Early on [in my illness], there was no way I could have traveled on my own [to a clinical trial]. It would have taken everything out of me,” says Kyle*, a Vermont filmmaker who has been living with Long COVID since April 2022. While the launch of the ambitious clinical trial was not without its kinks — Kyle and others say they experienced scheduling delays, such as receiving the drug/placebo or payments late, that prolonged the process — he says that PAX LC has opened a window for him that otherwise would have been unavailable. “It was nice the trial tried to make it work for people, understanding that part of our illness is incapacity.”
The PAX LC Trial was largely made possible by the Yale Center for Clinical Investigation team, with Yashira Henriquez, the clinical investigation project manager, who is based in New York, playing a central role. Her daily activities included monitoring participants’ surveys and checking in on anyone who had reported adverse events. “Since there are no study visits, we texted or called each other all the time,” she says. The most common adverse event was dysgeusia, or a metallic taste in the mouth, which has been a common complaint among patients with acute COVID who have received the drug. “In this type of trial where you never meet someone face-to-face, you need to show even more compassion to make them feel at ease.”
In previous roles, Henriquez had worked on studies evaluating treatment for acute COVID-19 infections but wasn’t as familiar with Long COVID. Her experience working for the PAX LC Trial has opened her eyes to the dire need for more research in treatments for the chronic condition.
“I’ve worked on clinical trials in numerous positions, but this is one where patients are especially in desperate need to find answers,” she says. She recalls an hour-long phone conversation she had with a participant who was struggling with depression and becoming a stay-at-home mom after Long COVID forced her to leave her job. “It’s hard to see people who were living a healthy life end up in a state where they’re now debilitated and can’t even get out of bed.”
Most participants she worked with responded well to the design of the trial, she adds. Their cooperation allowed the study to run smoothly despite many of them being states away. “We didn’t have to bother them to do the surveys or get their study visits done,” Henriquez says, “because they all wanted to help.”
The “record rate” of enrollment speaks to the success of PAX LC Trial’s design. It can take years for clinical trials to finish the enrollment period. PAX LC finished in under a year. Although it is a small trial of 100 participants, given its numerous restrictions around eligibility to participate and requirements for collecting blood and saliva specimens for Iwasaki’s lab, enrollment would have taken significantly longer without its decentralized format, Krumholz says.
Previously, a trial like this was not practical. But now, the evolution of new technologies is making the possibility of decentralized trials more accessible than ever. Recently, for example, Krumholz co-founded Hugo Health, a platform that enables people to transfer their medical records more seamlessly, with their permission, to researchers. “Because of the 21 st century Cures Act [ a 2016 law created to advance medical product innovation], people can get access to their medical records and share them with trusted partners,” he says. “Now we don’t have to go through a Byzantine set of obstacles through a particular health system to access a patient’s medical data – we just need to work in partnership with participants.”
Traveling to each participant’s home would seem to be a pricey endeavor. But given its greater efficiency, the design of the PAX LC Trial has actually cut down on costs, Krumholz argues. If it had been run like a standard clinical trial, his team would have had to set up study sites across the country. The associated costs would have been extremely high; also, the complexities involved in a multitude of institutional review boards (IRBs) whose approval would be needed, would have significantly slowed the study’s progress. Furthermore, many clinical trials need to extend their study periods as they struggle to retain participants, which also adds to the costs.
“Despite the fact that our trial is sort of built as a concierge service for each individual patient, it was much less expensive than what we had been doing,” Krumholz says. “What I hope will happen is that it will be a competitive advantage to do decentralized trials.”
About a month into the PAX LC Trial, Ezra noticed that his symptoms had dramatically declined. He still doesn’t know if he had taken the drug or the placebo, but “the brain fog is gone, the memory issues are gone, the cognitive problems seem to have pretty much evaporated,” he says. He can take a 45-minute walk again and was able to return to work. He is not cured, however. He still has symptoms of postural orthostatic tachycardia syndrome (POTS) [a disorder in which standing up triggers symptoms such as rapid heart rate] and gets winded easily.
Unfortunately, for many people with Long COVID, there won’t be a magic pill that restores them to normal. More clinical trials aimed at other possible underlying mechanisms of Long COVID, such as autoimmune dysfunction, will be needed. And these patients, many of whom are struggling to work and provide for their families, don’t have years to wait. Decentralized trials could help bring much-needed answers more quickly.
And this type of trial doesn’t need to be confined to Long COVID. “I’m confident the way we set this up can work with almost any population,” says Krumholz. He is optimistic that PAX LC will inspire other researchers to adopt a more participant-centric format for different kinds of research.
“Our trial is like Kitty Hawk [the Wright Flyer], we just needed to show that it could fly,” he says. “The hope is that soon there will be fleets of planes that adopt this approach.”
Root out friction in every digital experience, super-charge conversion rates, and optimise digital self-service
Uncover insights from any interaction, deliver AI-powered agent coaching, and reduce cost to serve
Increase revenue and loyalty with real-time insights and recommendations delivered straight to teams on the ground
Know how your people feel and empower managers to improve employee engagement, productivity, and retention
Take action in the moments that matter most along the employee journey and drive bottom line growth
Whatever they’re are saying, wherever they’re saying it, know exactly what’s going on with your people
Get faster, richer insights with qual and quant tools that make powerful market research available to everyone
Run concept tests, pricing studies, prototyping + more with fast, powerful studies designed by UX research experts
Track your brand performance 24/7 and act quickly to respond to opportunities and challenges in your market
Meet the operating system for experience management
Popular Use Cases
The annual gathering of the experience leaders at the world’s iconic brands building breakthrough business results.
What is survey research.
15 min read Find out everything you need to know about survey research, from what it is and how it works to the different methods and tools you can use to ensure you’re successful.
Survey research is the process of collecting data from a predefined group (e.g. customers or potential customers) with the ultimate goal of uncovering insights about your products, services, or brand overall .
As a quantitative data collection method, survey research can provide you with a goldmine of information that can inform crucial business and product decisions. But survey research needs careful planning and execution to get the results you want.
So if you’re thinking about using surveys to carry out research, read on.
Get started with our free survey software
Calling these methods ‘survey research’ slightly underplays the complexity of this type of information gathering. From the expertise required to carry out each activity to the analysis of the data and its eventual application, a considerable amount of effort is required.
As for how you can carry out your research, there are several options to choose from — face-to-face interviews, telephone surveys, focus groups (though more interviews than surveys), online surveys, and panel surveys.
Typically, the survey method you choose will largely be guided by who you want to survey, the size of your sample , your budget, and the type of information you’re hoping to gather.
Here are a few of the most-used survey types:
Before technology made it possible to conduct research using online surveys, telephone, and mail were the most popular methods for survey research. However face-to-face interviews were considered the gold standard — the only reason they weren’t as popular was due to their highly prohibitive costs.
When it came to face-to-face interviews, organisations would use highly trained researchers who knew when to probe or follow up on vague or problematic answers. They also knew when to offer assistance to respondents when they seemed to be struggling. The result was that these interviewers could get sample members to participate and engage in surveys in the most effective way possible, leading to higher response rates and better quality data.
While phone surveys have been popular in the past, particularly for measuring general consumer behaviour or beliefs, response rates have been declining since the 1990s .
Phone surveys are usually conducted using a random dialling system and software that a researcher can use to record responses.
This method is beneficial when you want to survey a large population but don’t have the resources to conduct face-to-face research surveys or run focus groups, or want to ask multiple-choice and open-ended questions .
The downsides are they can: take a long time to complete depending on the response rate, and you may have to do a lot of cold-calling to get the information you need.
You also run the risk of respondents not being completely honest . Instead, they’ll answer your survey questions quickly just to get off the phone.
Focus groups are a separate qualitative methodology rather than surveys — even though they’re often bunched together. They’re normally used for survey pretesting and designing , but they’re also a great way to generate opinions and data from a diverse range of people.
Focus groups involve putting a cohort of demographically or socially diverse people in a room with a moderator and engaging them in a discussion on a particular topic, such as your product, brand, or service.
They remain a highly popular method for market research , but they’re expensive and require a lot of administration to conduct and analyse the data properly.
You also run the risk of more dominant members of the group taking over the discussion and swaying the opinions of other people — potentially providing you with unreliable data.
Online surveys have become one of the most popular survey methods due to being cost-effective, enabling researchers to accurately survey a large population quickly.
Online surveys can essentially be used by anyone for any research purpose – we’ve all seen the increasing popularity of polls on social media (although these are not scientific).
Using an online survey allows you to ask a series of different question types and collect data instantly that’s easy to analyse with the right software.
There are also several methods for running and distributing online surveys that allow you to get your questionnaire in front of a large population at a fraction of the cost of face-to-face interviews or focus groups.
This is particularly true when it comes to mobile surveys as most people with a smartphone can access them online.
However, you have to be aware of the potential dangers of using online surveys, particularly when it comes to the survey respondents. The biggest risk is because online surveys require access to a computer or mobile device to complete, they could exclude elderly members of the population who don’t have access to the technology — or don’t know how to use it.
It could also exclude those from poorer socio-economic backgrounds who can’t afford a computer or consistent internet access. This could mean the data collected is more biased towards a certain group and can lead to less accurate data when you’re looking for a representative population sample.
When it comes to surveys, every voice matters.
A panel survey involves recruiting respondents who have specifically signed up to answer questionnaires and who are put on a list by a research company. This could be a workforce of a small company or a major subset of a national population. Usually, these groups are carefully selected so that they represent a sample of your target population — giving you balance across criteria such as age, gender, background, and so on.
Panel surveys give you access to the respondents you need and are usually provided by the research company in question. As a result, it’s much easier to get access to the right audiences as you just need to tell the research company your criteria. They’ll then determine the right panels to use to answer your questionnaire.
However, there are downsides. The main one being that if the research company offers its panels incentives, e.g. discounts, coupons, money — respondents may answer a lot of questionnaires just for the benefits.
This might mean they rush through your survey without providing considered and truthful answers. As a consequence, this can damage the credibility of your data and potentially ruin your analyses.
Depending on the research method you use, there are lots of benefits to conducting survey research for data collection. Here, we cover a few:
Most research surveys are easy to set up, administer and analyse. As long as the planning and survey design is thorough and you target the right audience , the data collection is usually straightforward regardless of which survey type you use.
Survey research can be relatively cheap depending on the type of survey you use.
Generally, qualitative research methods that require access to people in person or over the phone are more expensive and require more administration.
Online surveys or mobile surveys are often more cost-effective for market research and can give you access to the global population for a fraction of the cost.
Again, depending on the type of survey, you can obtain survey results from an entire population at a relatively low price. You can also administer a large variety of survey types to fit the project you’re running.
Using survey software, you can use advanced statistical analysis techniques to gain insights into your responses immediately.
Analysis can be conducted using a variety of parameters to determine the validity and reliability of your survey data at scale.
While most people view surveys as a quantitative research method, they can just as easily be adapted to gain qualitative information by simply including open-ended questions or conducting interviews face to face.
While surveys are a great way to obtain data, that data on its own is useless unless it can be analysed and developed into actionable insights.
The easiest, and most effective way to measure survey results, is to use a dedicated research tool that puts all of your survey results into one place.
When it comes to survey measurement, there are four measurement types to be aware of that will determine how you treat your different survey results:
With a nominal scale , you can only keep track of how many respondents chose each option from a question, and which response generated the most selections.
An example of this would be simply asking a responder to choose a product or brand from a list.
You could find out which brand was chosen the most but have no insight as to why.
Ordinal scales are used to judge an order of preference. They do provide some level of quantitative value because you’re asking responders to choose a preference of one option over another.
Ratio scales can be used to judge the order and difference between responses. For example, asking respondents how much they spend on their weekly shopping on average.
In an interval scale, values are lined up in order with a meaningful difference between the two values — for example, measuring temperature or measuring a credit score between one value and another.
Conducting a survey and collecting data is relatively straightforward, but it does require some careful planning and design to ensure it results in reliable data.
What do you want to learn from the survey? How is the data going to help you? Having a hypothesis or series of assumptions about survey responses will allow you to create the right questions to test them.
Once you’ve got your hypotheses or assumptions, write out the questions you need answering to test your theories or beliefs. Be wary about framing questions that could lead respondents or inadvertently create biased responses .
Your survey should include a variety of question types and should aim to obtain quantitative data with some qualitative responses from open-ended questions. Using a mix of questions (simple Yes/ No, multiple-choice, rank in order, etc) not only increases the reliability of your data but also reduces survey fatigue and respondents simply answering questions quickly without thinking.
Before sending your questionnaire out, you should test it (e.g. have a random internal group do the survey) and carry out A/B tests to ensure you’ll gain accurate responses.
Depending on your objectives, you might want to target the general population with your survey or a specific segment of the population. Once you’ve narrowed down who you want to target, it’s time to send out the survey.
After you’ve deployed the survey, keep an eye on the response rate to ensure you’re getting the number you expected. If your response rate is low, you might need to send the survey out to a second group to obtain a large enough sample — or do some troubleshooting to work out why your response rates are so low. This could be down to your questions, delivery method, selected sample, or otherwise.
Once you’ve got your results back, it’s time for the fun part.
Break down your survey responses using the parameters you’ve set in your objectives and analyse the data to compare to your original assumptions. At this stage, a research tool or software can make the analysis a lot easier — and that’s somewhere Qualtrics can help.
Gaining feedback from customers and leads is critical for any business, data gathered from surveys can prove invaluable for understanding your products and your market position, and with survey software from Qualtrics , it couldn’t be easier.
Used by more than 13,000 brands and supporting more than 1 billion surveys a year, Qualtrics empowers everyone in your organisation to gather insights and take action. No coding required — and your data is housed in one system.
Get feedback from more than 125 sources on a single platform and view and measure your data in one place to create actionable insights and gain a deeper understanding of your target customers.
Automatically run complex text and statistical analysis to uncover exactly what your survey data is telling you, so you can react in real-time and make smarter decisions.
We can help you with survey management, too. From designing your survey and finding your target respondents to getting your survey in the field and reporting back on the results, we can help you every step of the way.
And for expert market researchers and survey designers, Qualtrics features custom programming to give you total flexibility over question types, survey design, embedded data, and other variables.
No matter what type of survey you want to run, what target audience you want to reach, or what assumptions you want to test or answers you want to uncover, we’ll help you design, deploy and analyse your survey with our team of experts.
Start your survey research today with Qualtrics
Thematic analysis 11 min read, post event survey questions 10 min read, choosing the best survey tools 16 min read, survey app 11 min read, close-ended questions 7 min read, survey vs questionnaire 12 min read, likert scales 14 min read, request demo.
Ready to learn more about Qualtrics?
IMAGES
VIDEO
COMMENTS
6. Write a null hypothesis. If your research involves statistical hypothesis testing, you will also have to write a null hypothesis. The null hypothesis is the default position that there is no association between the variables. The null hypothesis is written as H 0, while the alternative hypothesis is H 1 or H a.
The Role of QuestionPro to Develop a Good Research Hypothesis. QuestionPro is a survey and research platform that provides tools for creating, distributing, and analyzing surveys. It plays a crucial role in the research process, especially when you're in the initial stages of hypothesis development. Here's how QuestionPro can help you to ...
It seeks to explore and understand a particular aspect of the research subject. In contrast, a research hypothesis is a specific statement or prediction that suggests an expected relationship between variables. It is formulated based on existing knowledge or theories and guides the research design and data analysis. 7.
INTRODUCTION. Scientific research is usually initiated by posing evidenced-based research questions which are then explicitly restated as hypotheses.1,2 The hypotheses provide directions to guide the study, solutions, explanations, and expected results.3,4 Both research questions and hypotheses are essentially formulated based on conventional theories and real-world processes, which allow the ...
3. Simple hypothesis. A simple hypothesis is a statement made to reflect the relation between exactly two variables. One independent and one dependent. Consider the example, "Smoking is a prominent cause of lung cancer." The dependent variable, lung cancer, is dependent on the independent variable, smoking. 4.
Step 5: Phrase your hypothesis in three ways. To identify the variables, you can write a simple prediction in if … then form. The first part of the sentence states the independent variable and the second part states the dependent variable. If a first-year student starts attending more lectures, then their exam scores will improve.
Survey research is defined as "the collection of information from a sample of individuals through their responses to questions" ( Check & Schutt, 2012, p. 160 ). This type of research allows for a variety of methods to recruit participants, collect data, and utilize various methods of instrumentation. Survey research can use quantitative ...
An effective hypothesis in research is clearly and concisely written, and any terms or definitions clarified and defined. Specific language must also be used to avoid any generalities or assumptions. Use the following points as a checklist to evaluate the effectiveness of your research hypothesis: Predicts the relationship and outcome.
Survey research can be relatively cheap depending on the type of survey you use. Generally, qualitative research methods that require access to people in person or over the phone are more expensive and require more administration. ... Having a hypothesis or series of assumptions about survey responses will allow you to create the right ...
A research hypothesis (also called a scientific hypothesis) is a statement about the expected outcome of a study (for example, a dissertation or thesis). To constitute a quality hypothesis, the statement needs to have three attributes - specificity, clarity and testability. Let's take a look at these more closely.
The steps to write a research hypothesis are: 1. Stating the problem: Ensure that the hypothesis defines the research problem. 2. Writing a hypothesis as an 'if-then' statement: Include the action and the expected outcome of your study by following a 'if-then' structure. 3.
There are 5 main steps in hypothesis testing: State your research hypothesis as a null hypothesis and alternate hypothesis (H o) and (H a or H 1 ). Collect data in a way designed to test the hypothesis. Perform an appropriate statistical test. Decide whether to reject or fail to reject your null hypothesis. Present the findings in your results ...
We will return to this hypothesis with a new way to answer it when we introduce correlational designs. We begin this chapter with an introduction to the research design that was illustrated here: the survey research design. 8.1 An Overview of Survey Designs A nonexperimental research design used to describe an individual or a group by having
Survey research means collecting information about a group of people by asking them questions and analyzing the results. To conduct an effective survey, follow these six steps: Determine who will participate in the survey. Decide the type of survey (mail, online, or in-person) Design the survey questions and layout.
A hypothesis can be defined as an assumption statement that is made on the basis of evidence so that this assumption can be tested to see if it might be true. It describes what you expect will happen in your research study before it has taken place and is therefore a prediction that you are trying to explore. Certain research studies may involve several hypotheses in cases where multiple ...
Encyclopedia of Survey Research Methods is a comprehensive reference work that covers all aspects of survey research, from design to analysis. Learn from experts in the field and access hundreds of entries on topics such as sampling, questionnaire design, measurement, data collection, and more.
A hypothesis is a tentative statement about the relationship between two or more variables. It is a specific, testable prediction about what you expect to happen in a study. It is a preliminary answer to your question that helps guide the research process. Consider a study designed to examine the relationship between sleep deprivation and test ...
A survey of 100 randomly sampled customers finds that 73 percent are very satisfied. To test the CEO's hypothesis, find the region of acceptance. Assume a significance level of 0.05. Solution: The analysis of survey data to test a hypothesis takes seven steps. We work through those steps below:
The survey is then constructed to test this model against observations of the phenomena. In contrast to survey research, a . survey. is simply a data collection tool for carrying out survey research. Pinsonneault and Kraemer (1993) defined a survey as a "means for gathering information about the characteristics, actions, or opinions of a ...
Survey research is a quantitative and qualitative method with two important characteristics. First, the variables of interest are measured using self-reports. In essence, survey researchers ask their participants (who are often called respondents in survey research) to report directly on their own thoughts, feelings, and behaviours.
Survey Research. Definition: Survey Research is a quantitative research method that involves collecting standardized data from a sample of individuals or groups through the use of structured questionnaires or interviews. The data collected is then analyzed statistically to identify patterns and relationships between variables, and to draw conclusions about the population being studied.
Hypothesis testing is a type of statistical testing that usually determines if there is enough evidence in the information gathered. When a hypothesis testing for surveys is conducted a person should be clear of the goals and objectives of the survey which simplifies the process of research and getting an accurate result.
None of these require hypotheses. But if you have a specific hypothesis, that can help in the design of the survey.in terms of what data need to be collected. Even when conducting analysis, there ...
This method involves giving a patient, client, or research study participant a survey or questionnaire to gauge their experience. The survey or questionnaire often asks about how someone feels or ...
Hypothesis 1: Ceteris paribus ... The Institute's primary aim is to assist the SMG through survey research in developing effective policies to improve the quality of life for its citizens. As part of this effort, Seoul Survey has been designed and implemented annually since 2003. The 2018 version of SS sampled all members (ages 15 and over ...
This study supports the hypothesis that distinct constellations of 15 interpretable judgment variables, along with contextual variables, could yield an efficient and highly scalable system for ...
Krumholz, Iwasaki, and their team designed the Pax LC Trial to test the persistent virus hypothesis. Cindy initiated the prescreening process by filling out a survey about pre-existing conditions and syncing her medical records to an app on her phone. Several months later, she received an email inviting her to move forward.
Survey research is the process of collecting data from a predefined group (e.g. customers or potential customers) with the ultimate goal of uncovering insights about your products, services, or brand overall. As a quantitative data collection method, survey research can provide you with a goldmine of information that can inform crucial business and product decisions.
The World Bank Enterprise Survey, conducted between 2014 and 2016, collected information from 1,184 manufacturing firms used in the research on structural reconfiguration. The hierarchical regression model helps to create prediction equations to test structural reconfiguration as the mediators of organizational efficiency and business ...