A .gov website belongs to an official government organization in the United States.
A lock ( ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.
- CDC Field Epidemiology Manual Chapters
Collecting and Analyzing Qualitative Data
At a glance.
Chapter 10 of The CDC Field Epidemiology Manual
Introduction
Qualitative research methods are a key component of field epidemiologic investigations because they can provide insight into the perceptions, values, opinions, and community norms where investigations are being conducted 1 2 . Open-ended inquiry methods, the mainstay of qualitative interview techniques, are essential in formative research for exploring contextual factors and rationales for risk behaviors that do not fit neatly into predefined categories. For example, during the 2014–2015 Ebola virus disease outbreaks in parts of West Africa, understanding the cultural implications of burial practices within different communities was crucial to designing and monitoring interventions for safe burials (see below). In program evaluations, qualitative methods can assist the investigator in diagnosing what went right or wrong as part of a process evaluation or in troubleshooting why a program might not be working as well as expected. When designing an intervention, qualitative methods can be useful in exploring dimensions of acceptability to increase the chances of intervention acceptance and success. When performed in conjunction with quantitative studies, qualitative methods can help the investigator confirm, challenge, or deepen the validity of conclusions than either component might have yielded alone 1 2 .
Qualitative Research During the Ebola Virus Disease Outbreaks in Parts of West Africa (2014)
Qualitative research was used extensively in response to the Ebola virus disease outbreaks in parts of West Africa to understand burial practices and to design culturally appropriate strategies to ensure safe burials. Qualitative studies were also used to monitor key aspects of the response.
In October 2014, Liberia experienced an abrupt and steady decrease in case counts and deaths in contrast with predicted disease models of an increased case count. At the time, communities were resistant to entering Ebola treatment centers, raising the possibility that patients were not being referred for care and communities might be conducting occult burials.
To assess what was happening at the community level, the Liberian Emergency Operations Center recruited epidemiologists from the US Department of Health and Human Services/Centers for Disease Control and Prevention and the African Union to investigate the problem.
Teams conducted in-depth interviews and focus group discussions with community leaders, local funeral directors, and coffin makers and learned that communities were not conducting occult burials and that the overall number of burials was less than what they had experienced in previous years. Other key findings included the willingness of funeral directors to cooperate with disease response efforts, the need for training of funeral home workers, and considerable community resistance to cremation practices. These findings prompted the Emergency Operations Center to open a burial ground for Ebola decedents, support enhanced testing of burials in the private sector, and train private-sector funeral workers regarding safe burial practices.
Source: Melissa Corkum, personal communication
Choosing When to Apply Qualitative Methods
Similar to quantitative approaches, qualitative research seeks answers to specific questions by using rigorous approaches to collecting and compiling information and producing findings that can be applicable beyond the study population. The fundamental difference in approaches lies in how they translate real-life complexities of initial observations into units of analysis. Data collected in qualitative studies typically are in the form of text or visual images, which provide rich sources of insight but also tend to be bulky and time-consuming to code and analyze. Practically speaking, qualitative study designs tend to favor small, purposively selected samples 1 ideal for case studies or in-depth analysis. The combination of purposive sampling and open-ended question formats deprive qualitative study designs of the power to quantify and generalize conclusions, one of the key limitations of this approach.
Qualitative scientists might argue, however, that the generalizability and precision possible through probabilistic sampling and categorical outcomes are achieved at the cost of enhanced validity, nuance, and naturalism that less structured approaches offer 3 . Open-ended techniques are particularly useful for understanding subjective meanings and motivations underlying behavior. They enable investigators to be equally adept at exploring factors observed and unobserved, intentions as well as actions, internal meanings as well as external consequences, options considered but not taken, and unmeasurable as well as measurable outcomes. These methods are important when the source of or solution to a public health problem is rooted in local perceptions rather than objectively measurable characteristics selected by outside observers 3 . Ultimately, such approaches have the ability to go beyond quantifying questions of how much or how many to take on questions of how or why from the perspective and in the words of the study subjects themselves 1 2 .
Another key advantage of qualitative methods for field investigations is their flexibility 4 . Qualitative designs not only enable but also encourage flexibility in the content and flow of questions to challenge and probe for deeper meanings or follow new leads if they lead to deeper understanding of an issue 5 . It is not uncommon for topic guides to be adjusted in the course of fieldwork to investigate emerging themes relevant to answering the original study question. As discussed herein, qualitative study designs allow flexibility in sample size to accommodate the need for more or fewer interviews among particular groups to determine the root cause of an issue (see the section on Sampling and Recruitment in Qualitative Research). In the context of field investigations, such methods can be extremely useful for investigating complex or fast-moving situations where the dimensions of analysis cannot be fully anticipated.
Ultimately, the decision whether to include qualitative research in a particular field investigation depends mainly on the nature of the research question itself. Certain types of research topics lend themselves more naturally to qualitative rather than other approaches ( Table 10.1 ). These include exploratory investigations when not enough is known about a problem to formulate a hypothesis or develop a fixed set of questions and answer codes. They include research questions where intentions matter as much as actions and "why?" or "why not?" questions matter as much as precise estimation of measured outcomes. Qualitative approaches also work well when contextual influences, subjective meanings, stigma, or strong social desirability biases lower faith in the validity of responses coming from a relatively impersonal survey questionnaire interview.
The availability of personnel with training and experience in qualitative interviewing or observation is critical for obtaining the best quality data but is not absolutely required for rapid assessment in field settings. Qualitative interviewing requires a broader set of skills than survey interviewing. It is not enough to follow a topic guide like a questionnaire, in order, from top to bottom. A qualitative interviewer must exercise judgment to decide when to probe and when to move on, when to encourage, challenge, or follow relevant leads even if they are not written in the topic guide. Ability to engage with informants, connect ideas during the interview, and think on one's feet are common characteristics of good qualitative interviewers. By far the most important qualification in conducting qualitative fieldwork is a firm grasp of the research objectives; with this qualification, a member of the research team armed with curiosity and a topic guide can learn on the job with successful results.
Examples of research topics for which qualitative methods should be considered for field investigations
Research topic
Exploratory research
The relevant questions or answer options are unknown in advance
In-depth case studies Situation analyses by viewing a problem from multiple perspectives Hypothesis generation
Understanding the role of context
Risk exposure or care-seeking behavior is embedded in particular social or physical environments
Key barriers or enablers to effective response Competing concerns that might interfere with each other Environmental behavioral interactions
Understanding the role of perceptions and subjective meaning
Different perception or meaning of the same observable facts influence risk exposure or behavioral response
Why or why not questions Understanding how persons make health decisions Exploring options considered but not taken
Understanding context and meaning of hidden, sensitive, or illegal behaviors
Legal barriers or social desirability biases prevent candid reporting by using conventional interviewing methods
Risky sexual or drug use behaviors Quality-of-care questions Questions that require a higher degree of trust between respondent and interviewer to obtain valid answers
Evaluating how interventions work in practice
Evaluating What went right or, more commonly, what went wrong with a public health response Process or outcome evaluations Who benefited in what way from what perceived change in practice
‘How’ questions Why interventions fail Unintended consequences of programs Patient–provider interactions
Commonly Used Qualitative Methods in Field Investigations
Semi-structured interviews.
Semi-structured interviews can be conducted with single participants (in-depth or individual key informants) or with groups (focus group discussions [FGDs] or key informant groups). These interviews follow a suggested topic guide rather than a fixed questionnaire format. Topic guides typically consist of a limited number (10-15) of broad, open-ended questions followed by bulleted points to facilitate optional probing. The conversational back-and-forth nature of a semi-structured format puts the researcher and researched (the interview participants) on more equal footing than allowed by more structured formats. Respondents, the term used in the case of quantitative questionnaire interviews, become informants in the case of individual semi-structured in-depth interviews (IDIs) or participants in the case of FGDs. Freedom to probe beyond initial responses enables interviewers to actively engage with the interviewee to seek clarity, openness, and depth by challenging informants to reach below layers of self-presentation and social desirability. In this respect, interviewing is sometimes compared with peeling an onion, with the first version of events accessible to the public, including survey interviewers, and deeper inner layers accessible to those who invest the time and effort to build rapport and gain trust. (The theory of the active interview suggests that all interviews involve staged social encounters where the interviewee is constantly assessing interviewer intentions and adjusting his or her responses accordingly 1 . Consequently good rapport is important for any type of interview. Survey formats give interviewers less freedom to divert from the preset script of questions and formal probes.)
Individual In-Depth Interviews and Key-Informant Interviews
The most common forms of individual semi-structured interviews are IDIs and key informant interviews (KIIs). IDIs are conducted among informants typically selected for first-hand experience (e.g., service users, participants, survivors) relevant to the research topic. These are typically conducted as one-on-one face-to-face interviews (two-on-one if translators are needed) to maximize rapport-building and confidentiality. KIIs are similar to IDIs but focus on individual persons with special knowledge or influence (e.g., community leaders or health authorities) that give them broader perspective or deeper insight into the topic area (See: Identifying Barriers and Solutions to Improved Healthcare Worker Practices in Egypt ). Whereas IDIs tend to focus on personal experiences, context, meaning, and implications for informants, KIIs tend to steer away from personal questions in favor of expert insights or community perspectives. IDIs enable flexible sampling strategies and represent the interviewing reference standard for confidentiality, rapport, richness, and contextual detail. However, IDIs are time-and labor-intensive to collect and analyze. Because confidentiality is not a concern in KIIs, these interviews might be conducted as individual or group interviews, as required for the topic area.
Focus Group Discussions and Group Key Informant Interviews
FGDs are semi-structured group interviews in which six to eight participants, homogeneous with respect to a shared experience, behavior, or demographic characteristic, are guided through a topic guide by a trained moderator 6 . (Advice on ideal group interview size varies. The principle is to convene a group large enough to foster an open, lively discussion of the topic, and small enough to ensure all participants stay fully engaged in the process.) Over the course of discussion, the moderator is expected to pose questions, foster group participation, and probe for clarity and depth. Long a staple of market research, focus groups have become a widely used social science technique with broad applications in public health, and they are especially popular as a rapid method for assessing community norms and shared perceptions.
Focus groups have certain useful advantages during field investigations. They are highly adaptable, inexpensive to arrange and conduct, and often enjoyable for participants. Group dynamics effectively tap into collective knowledge and experience to serve as a proxy informant for the community as a whole. They are also capable of recreating a microcosm of social norms where social, moral, and emotional dimensions of topics are allowed to emerge. Skilled moderators can also exploit the tendency of small groups to seek consensus to bring out disagreements that the participants will work to resolve in a way that can lead to deeper understanding. There are also limitations on focus group methods. Lack of confidentiality during group interviews means they should not be used to explore personal experiences of a sensitive nature on ethical grounds. Participants may take it on themselves to volunteer such information, but moderators are generally encouraged to steer the conversation back to general observations to avoid putting pressure on other participants to disclose in a similar way. Similarly, FGDs are subject by design to strong social desirability biases. Qualitative study designs using focus groups sometimes add individual interviews precisely to enable participants to describe personal experiences or personal views that would be difficult or inappropriate to share in a group setting. Focus groups run the risk of producing broad but shallow analyses of issues if groups reach comfortable but superficial consensus around complex topics. This weakness can be countered by training moderators to probe effectively and challenge any consensus that sounds too simplistic or contradictory with prior knowledge. However, FGDs are surprisingly robust against the influence of strongly opinionated participants, highly adaptable, and well suited to application in study designs where systematic comparisons across different groups are called for.
Like FGDs, group KIIs rely on positive chemistry and the stimulating effects of group discussion but aim to gather expert knowledge or oversight on a particular topic rather than lived experience of embedded social actors. Group KIIs have no minimum size requirements and can involve as few as two or three participants.
Identifying Barriers and Solutions to Improved Healthcare Worker Practices in Egypt
Egypt's National Infection Prevention and Control (IPC) program undertook qualitative research to gain an understanding of the contextual behaviors and motivations of healthcare workers in complying with IPC guidelines. The study was undertaken to guide the development of effective behavior change interventions in healthcare settings to improve IPC compliance.
Key informant interviews and focus group discussions were conducted in two governorates among cleaning staff, nursing staff, and physicians in different types of healthcare facilities. The findings highlighted social and cultural barriers to IPC compliance, enabling the IPC program to design responses. For example,
- Informants expressed difficulty in complying with IPC measures that forced them to act outside their normal roles in an ingrained hospital culture. Response: Role models and champions were introduced to help catalyze change.
- Informants described fatalistic attitudes that undermined energy and interest in modifying behavior. Response: Accordingly, interventions affirming institutional commitment to change while challenging fatalistic assumptions were developed.
- Informants did not perceive IPC as effective. Response: Trainings were amended to include scientific evidence justifying IPC practices.
- Informants perceived hygiene as something they took pride in and were judged on. Response: Public recognition of optimal IPC practice was introduced to tap into positive social desirability and professional pride in maintaining hygiene in the work environment.
Qualitative research identified sources of resistance to quality clinical practice in Egypt's healthcare settings and culturally appropriate responses to overcome that resistance.
Source: Anna Leena Lohiniva, personal communication.
Visualization Methods
Visualization methods have been developed as a way to enhance participation and empower interviewees relative to researchers during group data collection 7 . Visualization methods involve asking participants to engage in collective problem- solving of challenges expressed through group production of maps, diagrams, or other images. For example, participants from the community might be asked to sketch a map of their community and to highlight features of relevance to the research topic (e.g., access to health facilities or sites of risk concentrations). Body diagramming is another visualization tool in which community members are asked to depict how and where a health threat affects the human body as a way of understanding folk conceptions of health, disease, treatment, and prevention. Ensuing debate and dialogue regarding construction of images can be recorded and analyzed in conjunction with the visual image itself. Visualization exercises were initially designed to accommodate groups the size of entire communities, but they can work equally well with smaller groups corresponding to the size of FGDs or group KIIs.
Sampling and Recruitment for Qualitative Research
Selecting a sample of study participants.
Fundamental differences between qualitative and quantitative approaches to research emerge most clearly in the practice of sampling and recruitment of study participants. Qualitative samples are typically small and purposive. In-depth interview informants are usually selected on the basis of unique characteristics or personal experiences that make them exemplary for the study, if not typical in other respects. Key informants are selected for their unique knowledge or influence in the study domain. Focus group mobilization often seeks participants who are typical with respect to others in the community having similar exposure or shared characteristics. Often, however, participants in qualitative studies are selected because they are exceptional rather than simply representative. Their value lies not in their generalizability but in their ability to generate insight into the key questions driving the study.
Determining Sample Size
Sample size determination for qualitative studies also follows a different logic than that used for probability sample surveys. For example, whereas some qualitative methods specify ideal ranges of participants that constitute a valid observation (e.g., focus groups), there are no rules on how many observations it takes to attain valid results. In theory, sample size in qualitative designs should be determined by the saturation principle , where interviews are conducted until additional interviews yield no additional insights into the topic of research 8 . Practically speaking, designing a study with a range in number of interviews is advisable for providing a level of flexibility if additional interviews are needed to reach clear conclusions.
Recruiting Study Participants
Recruitment strategies for qualitative studies typically involve some degree of participant self-selection (e.g., advertising in public spaces for interested participants) and purposive selection (e.g., identification of key informants). Purposive selection in community settings often requires authorization from local authorities and assistance from local mobilizers before the informed consent process can begin. Clearly specifying eligibility criteria is crucial for minimizing the tendency of study mobilizers to apply their own filters regarding who reflects the community in the best light. In addition to formal eligibility criteria, character traits (e.g., articulate and interested in participating) and convenience (e.g., not too far away) are legitimate considerations for whom to include in the sample. Accommodations to personality and convenience help to ensure the small number of interviews in a typical qualitative design yields maximum value for minimum investment. This is one reason why random sampling of qualitative informants is not only unnecessary but also potentially counterproductive.
Managing, Condensing, Displaying, and Interpreting Qualitative Data
Analysis of qualitative data can be divided into four stages 9 : data management, data condensation, data display, and drawing and verifying conclusions.
Managing Qualitative Data
From the outset, developing a clear organization system for qualitative data is important. Ideally, naming conventions for original data files and subsequent analysis should be recorded in a data dictionary file that includes dates, locations, defining individual or group characteristics, interviewer characteristics, and other defining features. Digital recordings of interviews or visualization products should be reviewed to ensure fidelity of analyzed data to original observations. If ethics agreements require that no names or identifying characteristics be recorded, all individual names must be removed from final transcriptions before analysis begins. If data are analyzed by using textual data analysis software, maintaining careful version control over the data files is crucial, especially when multiple coders are involved.
Condensing Qualitative Data
Condensing refers to the process of selecting, focusing, simplifying, and abstracting the data available at the time of the original observation, then transforming the condensed data into a data set that can be analyzed. In qualitative research, most of the time investment required to complete a study comes after the fieldwork is complete. A single hour of taped individual interview can take a full day to transcribe and additional time to translate if necessary. Group interviews can take even longer because of the difficulty of transcribing active group input. Each stage of data condensation involves multiple decisions that require clear rules and close supervision. A typical challenge is finding the right balance between fidelity to the rhythm and texture of original language and clarity of the translated version in the language of analysis. For example, discussions among groups with little or no education should not emerge after the transcription (and translation) process sounding like university graduates. Judgment must be exercised about which terms should be translated and which terms should be kept in vernacular because there is no appropriate term in English to capture the richness of its meaning.
Displaying Qualitative Data
After the initial condensation, qualitative analysis depends on how the data are displayed. Decisions regarding how data are summarized and laid out to facilitate comparison influence the depth and detail of the investigation's conclusions. Displays might range from full verbatim transcripts of interviews to bulleted summaries or distilled summaries of interview notes. In a field setting, a useful and commonly used display format is an overview chart in which key themes or research questions are listed in rows in a word processer table or in a spreadsheet and individual informant or group entry characteristics are listed across columns. Overview charts are useful because they allow easy, systematic comparison of results.
Drawing and Verifying Conclusions
Analyzing qualitative data is an iterative and ideally interactive process that leads to rigorous and systematic interpretation of textual or visual data. At least four common steps are involved:
- Reading and rereading. The core of qualitative analysis is careful, systematic, and repeated reading of text to identify consistent themes and interconnections emerging from the data. The act of repeated reading inevitably yields new themes, connections, and deeper meanings from the first reading. Reading the full text of interviews multiple times before subdividing according to coded themes is key to appreciating the full context and flow of each interview before subdividing and extracting coded sections of text for separate analysis.
- Coding. A common technique in qualitative analysis involves developing codes for labeling sections of text for selective retrieval in later stages of analysis and verification. Different approaches can be used for textual coding. One approach, structural coding , follows the structure of the interview guide. Another approach, thematic coding , labels common themes that appear across interviews, whether by design of the topic guide or emerging themes assigned based on further analysis. To avoid the problem of shift and drift in codes across time or multiple coders, qualitative investigators should develop a standard codebook with written definitions and rules about when codes should start and stop. Coding is also an iterative process in which new codes that emerge from repeated reading are layered on top of existing codes. Development and refinement of the codebook is inseparably part of the analysis.
- Analyzing and writing memos. As codes are being developed and refined, answers to the original research question should begin to emerge. Coding can facilitate that process through selective text retrieval during which similarities within and between coding categories can be extracted and compared systematically. Because no p values can be derived in qualitative analyses to mark the transition from tentative to firm conclusions, standard practice is to write memos to record evolving insights and emerging patterns in the data and how they relate to the original research questions. Writing memos is intended to catalyze further thinking about the data, thus initiating new connections that can lead to further coding and deeper understanding.
- Verifying conclusions. Analysis rigor depends as much on the thoroughness of the cross-examination and attempt to find alternative conclusions as on the quality of original conclusions. Cross-examining conclusions can occur in different ways. One way is encouraging regular interaction between analysts to challenge conclusions and pose alternative explanations for the same data. Another way is quizzing the data (i.e., retrieving coded segments by using Boolean logic to systematically compare code contents where they overlap with other codes or informant characteristics). If alternative explanations for initial conclusions are more difficult to justify, confidence in those conclusions is strengthened.
Coding and Analysis Requirements
Above all, qualitative data analysis requires sufficient time and immersion in the data. Computer textual software programs can facilitate selective text retrieval and quizzing the data, but discerning patterns and arriving at conclusions can be done only by the analysts. This requirement involves intensive reading and rereading, developing codebooks and coding, discussing and debating, revising codebooks, and recoding as needed until clear patterns emerge from the data. Although quality and depth of analysis is usually proportional to the time invested, a number of techniques, including some mentioned earlier, can be used to expedite analysis under field conditions.
- Detailed notes instead of full transcriptions. Assigning one or two note-takers to an interview can be considered where the time needed for full transcription and translation is not feasible. Even if plans are in place for full transcriptions after fieldwork, asking note-takers to submit organized summary notes is a useful technique for getting real-time feedback on interview content and making adjustments to topic guides or interviewer training as needed.
- Summary overview charts for thematic coding. (See discussion under "Displaying Data.") If there is limited time for full transcription and/or systematic coding of text interviews using textual analysis software in the field, an overview chart is a useful technique for rapid manual coding.
- Thematic extract files. This is a slightly expanded version of manual thematic coding that is useful when full transcriptions of interviews are available. With use of a word processing program, files can be sectioned according to themes, or separate files can be created for each theme. Relevant extracts from transcripts or analyst notes can be copied and pasted into files or sections of files corresponding to each theme. This is particularly useful for storing appropriate quotes that can be used to illustrate thematic conclusions in final reports or manuscripts.
- Teamwork. Qualitative analysis can be performed by a single analyst, but it is usually beneficial to involve more than one. Qualitative conclusions involve subjective judgment calls. Having more than one coder or analyst working on a project enables more interactive discussion and debate before reaching consensus on conclusions.
- Systematic coding.
- Selective retrieval of coded segments.
- Verifying conclusions ("quizzing the data").
- Working on larger data sets with multiple separate files.
- Working in teams with multiple coders to allow intercoder reliability to be measured and monitored.
The most widely used software packages (e.g., NVivo [QSR International Pty. Ltd., Melbourne, VIC, Australia] and ATLAS.ti [Scientific Software Development GmbH, Berlin, Germany]) evolved to include sophisticated analytic features covering a wide array of applications but are relatively expensive in terms of license cost and initial investment in time and training. A promising development is the advent of free or low-cost Web-based services (e.g., Dedoose [Sociocultural Research Consultants LLC, Manhattan Beach, CA]) that have many of the same analytic features on a more affordable subscription basis and that enable local research counterparts to remain engaged through the analysis phase (see Teamwork criteria). The start-up costs of computer-assisted analysis need to be weighed against their analytic benefits, which tend to decline with the volume and complexity of data to be analyzed. For rapid situational analyses or small scale qualitative studies (e.g. fewer than 30 observations as an informal rule of thumb), manual coding and analysis using word processing or spreadsheet programs is faster and sufficient to enable rigorous analysis and verification of conclusions.
Qualitative methods belong to a branch of social science inquiry that emphasizes the importance of context, subjective meanings, and motivations in understanding human behavior patterns. Qualitative approaches definitionally rely on open-ended, semistructured, non-numeric strategies for asking questions and recording responses. Conclusions are drawn from systematic visual or textual analysis involving repeated reading, coding, and organizing information into structured and emerging themes. Because textual analysis is relatively time-and skill-intensive, qualitative samples tend to be small and purposively selected to yield the maximum amount of information from the minimum amount of data collection. Although qualitative approaches cannot provide representative or generalizable findings in a statistical sense, they can offer an unparalleled level of detail, nuance, and naturalistic insight into the chosen subject of study. Qualitative methods enable investigators to “hear the voice” of the researched in a way that questionnaire methods, even with the occasional open-ended response option, cannot.
Whether or when to use qualitative methods in field epidemiology studies ultimately depends on the nature of the public health question to be answered. Qualitative approaches make sense when a study question about behavior patterns or program performance leads with why, why not , or how . Similarly, they are appropriate when the answer to the study question depends on understanding the problem from the perspective of social actors in real-life settings or when the object of study cannot be adequately captured, quantified, or categorized through a battery of closed-ended survey questions (e.g., stigma or the foundation of health beliefs). Another justification for qualitative methods occurs when the topic is especially sensitive or subject to strong social desirability biases that require developing trust with the informant and persistent probing to reach the truth. Finally, qualitative methods make sense when the study question is exploratory in nature, where this approach enables the investigator the freedom and flexibility to adjust topic guides and probe beyond the original topic guides.
Given that the conditions just described probably apply more often than not in everyday field epidemiology, it might be surprising that such approaches are not incorporated more routinely into standard epidemiologic training. Part of the answer might have to do with the subjective element in qualitative sampling and analysis that seems at odds with core scientific values of objectivity. Part of it might have to do with the skill requirements for good qualitative interviewing, which are generally more difficult to find than those required for routine survey interviewing.
For the field epidemiologist unfamiliar with qualitative study design, it is important to emphasize that obtaining important insights from applying basic approaches is possible, even without a seasoned team of qualitative researchers on hand to do the work. The flexibility of qualitative methods also tends to make them forgiving with practice and persistence. Beyond the required study approvals and ethical clearances, the basic essential requirements for collecting qualitative data in field settings start with an interviewer having a strong command of the research question, basic interactive and language skills, and a healthy sense of curiosity, armed with a simple open-ended topic guide and a tape recorder or note-taker to capture the key points of the discussion. Readily available manuals on qualitative study design, methods, and analysis can provide additional guidance to improve the quality of data collection and analysis.
- Patton MQ. Qualitative research and evaluation methods: integrating theory and practice . 4th ed. Thousand Oaks, CA: Sage; 2015.
- Hennink M, Hutter I, Bailey A. Qualitative research methods . Thousand Oaks, CA: Sage; 2010.
- Lincoln YS, Guba EG. The constructivist credo . Walnut Creek, CA: Left Coast Press; 2013.
- Mack N, Woodsong C, MacQueen KM, Guest G, Namey E. Qualitative research methods: a data collectors field guide. https://www.fhi360.org/sites/default/files/media/documents/Qualitative%20Research%20Methods%20-%20A%20Data%20Collector%27s%20Field%20Guide.pdf
- Kvale S, Brinkmann S. Interviews: learning the craft of qualitative research . Thousand Oaks, CA: Sage; 2009:230–43.
- Krueger RA, Casey MA. Focus groups: a practical guide for applied research . Thousand Oaks, CA: Sage; 2014.
- Margolis E, Pauwels L. The Sage handbook of visual research methods . Thousand Oaks, CA: Sage; 2011.
- Mason M. Sample size and saturation in PhD studies using qualitative interviews. Forum : Qualitative Social Research/Sozialforschung. 2010;11(3).
- Miles MB, Huberman AM, Saldana J. Qualitative data analysis: a methods sourcebook . 3rd ed. Thousand Oaks, CA: Sage; 2014.
- Silver C, Lewins A. Using software in qualitative research: a step-by-step guide . Thousand Oaks, CA; Sage: 2014.
Field Epi Manual
The CDC Field Epidemiology Manual is a definitive guide to investigating acute public health events on the ground and in real time.
An official website of the United States government
Official websites use .gov A .gov website belongs to an official government organization in the United States.
Secure .gov websites use HTTPS A lock ( Lock Locked padlock icon ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.
- Publications
- Account settings
- Advanced Search
- Journal List
A review of qualitative data analysis practices in health education and health behavior research
Ilana g raskind , msc, rachel c shelton , scd, mph, dawn l comeau , phd, hannah l f cooper , scd, derek m griffith , phd, michelle c kegler , drph, mph.
- Author information
- Article notes
- Copyright and License information
Corresponding author.
Issue date 2019 Feb.
Data analysis is one of the most important, yet least understood stages of the qualitative research process. Through rigorous analysis, data can illuminate the complexity of human behavior, inform interventions, and give voice to people’s lived experiences. While significant progress has been made in advancing the rigor of qualitative analysis, the process often remains nebulous. To better understand how our field conducts and reports qualitative analysis, we reviewed qualitative papers published in Health Education & Behavior between 2000–2015. Two independent reviewers abstracted information in the following categories: data management software, coding approach, analytic approach, indicators of trustworthiness, and reflexivity. Of the forty-eight (n=48) articles identified, the majority (n=31) reported using qualitative software to manage data. Double-coding transcripts was the most common coding method (n=23); however, nearly one-third of articles did not clearly describe the coding approach. Although terminology used to describe the analytic process varied widely, we identified four overarching trajectories common to most articles (n=37). Trajectories differed in their use of inductive and deductive coding approaches, formal coding templates, and rounds or levels of coding. Trajectories culminated in the iterative review of coded data to identify emergent themes. Few papers explicitly discussed trustworthiness or reflexivity. Member checks (n=9), triangulation of methods (n=8), and peer debriefing (n=7) were the most common. Variation in the type and depth of information provided poses challenges to assessing quality and enabling replication. Greater transparency and more intentional application of diverse analytic methods can advance the rigor and impact of qualitative research in our field.
Keywords: health behavior, health promotion, qualitative methods, research design, training health professionals
Introduction
Data analysis is one of the most powerful, yet least understood stages of the qualitative research process. During this phase extensive fieldwork and illustrative data are transformed into substantive and actionable conclusions. In the field of health education and health behavior, rigorous data analysis can elucidate the complexity of human behavior, facilitate the development and implementation of impactful programs and interventions, and give voice to the lived experiences of inequity. While tremendous progress has been made in advancing the rigor of qualitative analysis, persistent misconceptions that such methods can be intuited rather than intentionally applied, coupled with inconsistent and vague reporting, continue to obscure the process ( Miles, Huberman, & Saldana, 2014 ). In an era of public health grounded in evidence-based research and practice, rigorously conducting, documenting, and reporting qualitative analysis is critical for the generation of reliable and actionable knowledge.
There is no single “right” way to engage in qualitative analysis ( Saldana & Omasta, 2018 ). The guiding inquiry framework, research questions, participants, context, and type of data collected should all inform the choice of analytic method ( Creswell & Poth, 2018 ; Saldana & Omasta, 2018 ). While the diversity and flexibility of methods for analysis may put the qualitative researcher in a more innovative position than their quantitative counterparts ( Miles et al., 2014 ), it also makes rigorous application and transparent reporting even more important ( Hennink, Hutter, & Bailey, 2011 ). Unlike many forms of quantitative analysis, qualitative analytic methods are far less likely to have standardized, widely agreed upon definitions and procedures ( Miles et al., 2014 ). The phrase thematic analysis , for example, may capture a variety of approaches and methodological tools, limiting the reader’s ability to accurately assess the rigor and credibility of the research. An explicit description of how data were condensed, patterns identified, and interpretations substantiated is likely of much greater use in assessing quality and facilitating replication. Yet, despite increased attention to the systematization of qualitative research ( Levitt et al., 2018 ; O’Brien et al., 2014 ; Tong, Sainsbury, & Craig, 2007 ), many studies remain vague in their reporting of how the researcher moved from “1,000 pages of field notes to the final conclusions” ( Miles et al., 2014 ).
Reflecting on the relevance of qualitative methods to the field of health education and health behavior, and challenges still facing the paradigm, we were interested in understanding how our field conducts and reports qualitative data analysis. In a companion paper ( Kegler et al. 2018 ), we describe our wider review of qualitative articles published in Health Education & Behavior ( HE&B ) from 2000 to 2015, broadly focused on how qualitative inquiry frameworks inform study design and study implementation. Upon conducting our initial review, we discovered that our method for abstracting information related to data analysis—documenting the labels researchers applied to analytic methods—shed little light on the concrete details of their analytic processes. As a result, we conducted a second round of review focused on how analytic approaches and techniques were applied. In particular, we assessed data preparation and management, approaches to data coding, analytic trajectories, methods for assessing credibility and trustworthiness, and approaches to reflexivity. Our objective was to develop a greater understanding of how our field engages in qualitative data analysis, and identify opportunities for strengthening our collective methodological toolbox.
Our methods are described in detail in a companion paper ( Kegler et al. 2018 ). Briefly, eligible articles were published in HE&B between 2000 and 2015 and used qualitative research methods. We excluded mixed methods studies because of differences in underlying paradigms, study design, and methods for analysis and interpretation. We reviewed 48 papers using an abstraction form designed to assess 10 main topics: qualitative inquiry framework, sampling strategy, data collection methods, data management software, coding approach, analytic approach, reporting of results, use of theory, indicators of trustworthiness, and reflexivity. The present paper reports results on data management software, coding approach, analytic approach, indicators of trustworthiness, and reflexivity.
Each article was initially double-coded by a team of six researchers, with one member of each coding pair reviewing the completed abstraction forms and noting discrepancies. This coder fixed discrepancies that could be easily resolved by re-reviewing the full text (e.g. sample size); a third coder reviewed more challenging discrepancies, which were then discussed with a member of the coding pair until consensus was reached. Data were entered into an Access database, and queries were generated to summarize results for each topic. Preliminary results were shared with all co-authors for review, and then discussed as a group.
New topics of interest emerged from the first round of review regarding how analytic approaches and techniques were applied. Two of the authors conducted a second round of review focused on: use of software, how authors discussed achieving coding consensus, use of matrices, analytic references cited, variation in how authors used the terms code and theme , and identification of common analytic trajectories, including how themes were identified, and the process of grouping themes or concepts. To facilitate the second round of review, the analysis section of each article was excerpted into a single document. One reviewer populated a spreadsheet with text from each article pertinent to the aforementioned categories, and summarized the content within each category. These results informed the development of a formal abstraction form. Two reviewers independently completed an abstraction form for each article’s analysis section and met to resolve discrepancies. For three of the categories (use of the terms code and theme; how themes were identified; and the process of grouping themes or concepts), we do not report counts or percentages because the level of detail provided was often insufficient to determine with certainty whether a particular strategy or combination of strategies was used.
Data preparation and management
We examined several dimensions of the data preparation and management process ( Table 1 ). The vast majority of papers (87.5%) used verbatim transcripts as the primary data source. Most others used detailed written summaries of interviews or a combination of transcripts and written summaries (14.6%). We documented whether qualitative software was mentioned and which packages were most commonly used. Fourteen of the articles used Atlas.ti (29.2%) and another seventeen (35.4%) did not report using software. NVivo and its predecessor NUD-IST were somewhat common (20.8%), and Ethnograph was used in two articles. Several other software packages were mentioned in one of the papers (e.g. AnSWR, EthnoNotes). Of those reporting use of a software package, the most common use, in addition to the implied management of data, was to code transcripts (33.3%). Approximately 10.4% described using the software to generate code reports, and 8.3% described using the software to calculate inter-rater reliability. Two articles (4.2%) described using the software to draft memos or data summaries. The remainder did not provide detail on how the software was used (16.7%).
Approaches to data preparation and management in qualitative papers, Health Education & Behavior 2000-2015 (n=48)
Note . Percentages may sum to >100 due to the use of multiple approaches
Data coding and analysis
Coding and consensus.
Double coding of all transcripts was most common by far (47.9%), although a significant proportion of papers did not discuss their approach to coding or the description provided was unclear (31.3%) ( Table 2 ). Among the remaining papers, approaches included a single coder with a second analyst reviewing the codes (8.3%), a single coder only (6.3%), and double coding of a portion of the transcripts with single coding of the rest (6.3%). A related issue is how consensus was achieved among coders. Approximately two-thirds (64.6%) of articles discussed their process for reaching consensus. Most described reaching consensus on definitions of codes or coding of text units through discussions (43.8%), while some mentioned the use of an additional reviewer to resolve discrepancies (8.3%).
Approaches to data coding and analysis in qualitative papers, Health Education & Behavior 2000-2015 (n=48)
Analytic approaches named by authors
As reported in our companion paper, thematic analysis (22.9%), content analysis (20.8%), and grounded theory (16.7%) were most commonly named analytic approaches. Approximately 43.8% named an approach that was not reported by other authors, including inductive analysis, immersion/crystallization, issue focused analysis, and editing qualitative methodology. Approximately 20% of the articles reported using matrices during analysis; most described using them to compare codes or themes across cases or demographic groups (14.6%).
We also examined which references authors cited to support their analytic approach. Although editions varied over time, the most commonly cited references included: Miles and Huberman (1984 , 1994 ); Bernard (1994 , 2002 , 2005 ), Bernard & Ryan (2010) , or Ryan & Bernard (2000 , 2003 ); Patton (1987 , 1990 , 1999 , 2002 ); and Strauss & Corbin (1994 , 1998 ) or Corbin & Strauss (1990) . These authors were cited in over five papers. Other references cited in 3–5 papers included: Lincoln and Guba (1985) or Guba (1978) ; Krueger (1994 , 1998 ) or Krueger & Casey (2000) ; Creswell (1998 , 2003 , 2007 ); and Charmaz (2000 , 2006 ).
Terminology: codes and themes
Given the diversity of definitions for the terms code and theme in the qualitative literature, we were interested in exploring how authors applied and distinguished the terms in their analyses. In over half of the articles, either both terms were not used, or the level of detail provided did not allow for clear categorization of how they were used. In the remainder of articles, we observed two general patterns: 1) the terms being used interchangeably and 2) themes emerging from codes.
Common analytic trajectories
In addition to examining various aspects of the analytic process as outlined above, we attempted to identify common overarching analytic trajectories or pathways. Authors generally used two approaches to indexing or breaking down and labeling the data (i.e., coding). The first approach (Pathways 1 and 2) was to create an initial list of codes based on theory, the literature, research questions, or the interview guide. The second approach (Pathways 3 and 4) was to read through transcripts to generate initial codes or patterns inductively. This approach was often labeled ‘open-coding’ or described as ‘making margin notes’ or ‘memoing’. We were unable to categorize 11 articles (22.9%) into one of the above pathways because the analysis followed a different trajectory (10.4%) or there was not enough detail reported (12.5%).
Those studies that began with initial or ‘start codes’ generally followed two pathways. The first (Pathway 1; 14.6%) was to code the data using the initial codes and then conduct a second round of coding within the ‘top level’ codes, often using open-coding to allow for identification of emergent themes. The second (Pathway 2; 18.8%) was to fully code the transcripts with the initial codes while simultaneously identifying emerging codes and modifying code definitions as needed. Those that did not start with an initial list of codes similarly followed two pathways. The first (Pathway 3; 33.3%) was to develop a formal coding template after open-coding (e.g., code transcripts in full with an iterative relabeling and creation of sub-codes) and the second (Pathway 4; 10.4%) was to use the initial codes generated from inductively reading the transcripts as the primary analytic step.
From all pathways, several approaches were used to identify themes: group discussions of salient themes, comparisons of coded data to develop or refine themes, combining related codes into themes, or extracting themes from codes. A small number of articles discussed or implied that themes or concepts were further grouped into broader categories or classes. However, the limited details provided by the authors made it difficult to ascertain the process used.
Validity, Trustworthiness, and Credibility
Few papers explicitly discussed techniques used to strengthen validity ( Table 3 ). Maxwell (1996) defines qualitative validity as “the correctness or credibility of a description, conclusion, explanation, interpretation, or other sort of account.” Member checks (18.8%; soliciting feedback on the credibility of the findings from members of the group from whom data were collected ( Creswell & Poth, 2018 )) and triangulation of methods (16.7%; assessing the consistency of findings across different data collection methods ( Patton, 2015 )) were the techniques reported most commonly. Peer debriefing (14.6%; external review of findings by a person familiar with the topic of study ( Creswell & Poth, 2018 )), prolonged engagement at a research site (10.4%), and analyst triangulation (10.4%; using multiple analysts to review and interpret findings ( Patton, 2015 )) were also reported. Triangulation of sources (assessing the consistency of findings across data sources within the same method ( Patton, 2015 )), audit trails (maintaining records of all steps taken throughout the research process to enable external auditing ( Creswell & Poth, 2018 )), and analysis of negative cases (examining cases that contradict or do not support emergent patterns and refining interpretations accordingly ( Creswell & Poth, 2018 )) were each mentioned only a few times. Lack of generalizability was discussed frequently, and was often a focus of the limitations section. Another commonly discussed threat to validity was an inability to draw conclusions about a construct or a domain of a construct because the sample was not diverse enough or because the number of participants in particular subgroups was too small. No papers discussed limitations to the completeness and accuracy of the data.
Approaches to establishing credibility, trustworthiness, and reflexivity in qualitative papers, Health Education & Behavior 2000-2015 (n=48)
Reflexivity
Reflexivity relates to the recognition that the perspective and position of the researcher shapes every step of the research process ( Creswell & Poth, 2018 ; Patton, 2015 ). Of the papers we reviewed, only four (8.3%) fully described the personal characteristics of the interviewers/facilitators (e.g. gender, occupation, training; Table 3 ). The majority (62.5%) provided minimal information about the interviewers (e.g. title or position), and 14 authors (29.2%) did not provide any information about personal characteristics. The vast majority of papers (87.5%) did not discuss the relationship and extent of interaction between interviewers/facilitators and participants. Only two papers explicitly discussed reflexivity, positionality, or potential personal bias based on the position of the researcher(s).
The present study sought to examine how the field of health education and health behavior has conducted and reported qualitative analysis over the past 15 years. We found great variation in the type and depth of analytic information reported. Although we were able to identify several broad analysis trajectories, the terminology used to describe the approaches varied widely, and the analytic techniques used were not described in great detail.
While the majority of articles reported whether data were double-coded, single-coded, or a combination thereof, additional detail on the coding method was infrequently provided. Saldaña (2016) describes two primary sets of coding methods that can be used in various combination: foundational first cycle codes (e.g. In Vivo, descriptive, open, structural), and conceptual second cycle codes (e.g. focused, pattern, theoretical). Each coding method possesses a unique set of strengths and can be used either solo or in tandem, depending upon the analytic objectives. For example, In Vivo codes, drawn verbatim from participant language and placed in quotes, are particularly useful for identifying and prioritizing participant voices and perspectives ( Saldana, 2016 ). Greater familiarity with, and more intentional application of, available techniques is likely to strengthen future research and accurately capture the ‘emic’ perspective of study participants.
Similarly, less than one quarter of studies described the use of matrices to organize coded data and support the identification of patterns, themes, and relationships. Matrices and other visual displays are widely discussed in the qualitative literature as an important organizing tool and stage in the analytic process ( Miles et al., 2014 ; Saldana & Omasta, 2018 ). They support the analyst in processing large quantities of data and drawing credible conclusions, tasks which are challenging for the brain to complete when the text is in extended form (i.e. coded transcripts) ( Miles et al., 2014 ). Like coding methods, myriad techniques exist for formulating matrices, which can be used for meeting various analytic objectives such as exploring specific variables, describing variability in findings, examining change across time, and explaining causal pathways ( Miles et al., 2014 ). Most qualitative software packages have extended capabilities in the construction of matrices and other visual displays.
Most authors reflected on their findings as a whole in article discussion sections. However, explicit descriptions of how themes or concepts were grouped together or related to one another—made into something greater than the sum of their parts—were rare. Miles et al. (2014) describe two tactics for systematically understanding the data as a whole: building a logical chain of evidence that describes how themes are causally linked to one another, and making conceptual coherence by aligning these themes with more generalized constructs that can be placed in a broader theoretical framework. Only one study in our review described the development of theory; while not a required outcome of analysis, moving from the identification of themes and patterns to such “higher-level abstraction” is what enables a study to transcend the particulars of the research project and draw more widely applicable conclusions ( Hennink et al., 2011 ; Saldana & Omasta, 2018 ).
All data analysis techniques will ideally flow from the broader inquiry framework and underlying paradigm within which the study is based ( Bradbury-Jones et al., 2017 ; Creswell & Poth, 2018 ; Patton, 2015 ). Yet, as reported in our companion paper ( Kegler et al. 2018 ), only six articles described the use of a well-established framework to guide their study (e.g. ethnography, grounded theory), making it difficult to assess how the reported analytic techniques aligned with the study’s broader assumptions and objectives. Interestingly, the most common analytic references were Miles & Huberman, Patton, and Bernard & Ryan, references which do not clearly align with a particular analytic approach or inquiry framework, and Strauss & Corbin, references aligned with grounded theory, an approach only reported in one of the included articles. In their Standards for Reporting Qualitative Research, O’Brien et al. (2014) assert that chosen methods should not only be described, but also justified. Encouraging intentional selection of an inquiry framework and complementary analytic techniques can strengthen qualitative research by compelling researchers to think through the implicit assumptions, limitations, and implications of their chosen approach.
When discussing validity of the research, papers overwhelmingly focused on the limited generalizability of their findings (a dimension of quantitative validity that Maxwell (1996) maintains is largely irrelevant for qualitative methods, yet one that is likely requested by peer reviewers and editors), and few discussed methods specific to qualitative research (e.g., member checks, reading for negative cases). It is notable that one of the least used strategies was the exploration of negative or disconfirming cases, rival explanations, and divergent patterns, given the importance of this approach in several foundational texts ( Miles et al., 2014 ; Patton, 2015 ). The primary focus on generalizability and the limited use of strategies designed to establish qualitative validity, may share a common root: the persistent hegemonic status of the quantitative paradigm. A more genuine embrace of qualitative methods in their own right may create space for a more comprehensive consideration of the specific nature of qualitative validity, and encourage investigators to apply and report such strategies in their work.
The researcher plays a unique role in qualitative inquiry: as the primary research instrument, they must subject their assumptions, decisions, actions, and conclusions to the same critical assessment they would any other instrument ( Hennink et al., 2011 ). However, we found that reflexivity and positionality on the part of the researcher was minimally addressed in the scope of the papers we reviewed. We encourage our fellow researchers to be more explicit in discussing how their training, position, sociodemographic characteristics, and relationship with participants may shape their own theoretical and methodological approach to the research, as well as their analysis and interpretation of findings. In some cases, this reflexivity may highlight the critical importance of building in efforts to enhance the credibility and trustworthiness of their research, including peer debriefs, audit trails, and member checks.
Limitations
The present study is subject to several important limitations. Clear consensus on qualitative reporting standards still does not exist, and it is not our intention to criticize the work of fellow researchers. Many of the articles included in our review were published prior to the release of Tong et al.’s (2007) Consolidated Criteria for Reporting Qualitative Research, O’Brien et al.’s (2014) Standards for Reporting Qualitative Research, and Levitt et al.’s (2018) Journal Article Reporting Standards for Qualitative Research. Further, we could only assess articles based on the information reported. The information included in the articles may be incomplete due to journal space limitations and may not reflect all analytic approaches and techniques used in the study. Finally, our review was restricted to articles published in HE&B and is not intended to represent the conduct and reporting of qualitative methods across the entire field of health education and health behavior, or public health more broadly. As an official journal of the Society for Public Health Education, we felt that HE&B would provide a high quality snapshot of the qualitative work being done in our field. Future reviews should include qualitative research published in other journals in the field.
Implications
Qualitative research is one of the most important tools we have for understanding the complexity of human behavior, including its context-specificity, multi-level determinants, cross-cultural meaning, and variation over time. Although no clear consensus exists on how to conduct and report qualitative analysis, thoughtful application and transparent reporting of key “analytic building blocks” may have at least four interconnected benefits: 1) spurring the use of a broader array of available methods; 2) improving the ability of readers and reviewers to critically appraise findings and contextualize them within the broader literature; 3) improving opportunities for replication; and 4) enhancing the rigor of qualitative research paradigms.
This effort may be aided by expanding the use of matrices and other visual displays, diverse methods for coding, and techniques for establishing qualitative validity, as well as greater attention to researcher positionality and reflexivity, the broader conceptual and theoretical frameworks that may emerge from analysis, and a decreased focus on generalizability as a limitation. Given the continued centrality of positivist research paradigms in the field of public health, supporting the use and reporting of uniquely qualitative methods and concepts must be the joint effort of researchers, practitioners, reviewers, and editors—an effort that is embedded within a broader endeavor to increase appreciation for the unique benefits of qualitative research.
Common analytic trajectories of qualitative papers in Health Education & Behavior , 2000–2015
Contributor Information
Ilana G. Raskind, Department of Behavioral Sciences and Health Education Rollins School of Public Health, Emory University 1518 Clifton Rd. NE, Atlanta, GA 30322, USA. [email protected].
Rachel C. Shelton, Columbia University, New York, NY, USA.
Dawn L. Comeau, Emory University, Atlanta, GA, USA.
Hannah L. F. Cooper, Emory University, Atlanta, GA, USA.
Derek M. Griffith, Vanderbilt University, Nashville, TN, USA.
Michelle C. Kegler, Emory University, Atlanta, GA, USA.
Bibliography and References Cited
- Bernard HR (1994). Research Methods in Anthropology: Qualitative and Quantitative Approaches (2nd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Bernard HR (2002). Research Methods in Anthropology: Qualitative and Quantitative Approaches (3rd ed.). Walnut Creek, CA: AltaMira Press. [ Google Scholar ]
- Bernard HR (2005). Research Methods in Anthropology: Qualitative and Quantitative Approaches (4th ed.). Walnut Creek, CA: AltaMira Press. [ Google Scholar ]
- Bernard HR, & Ryan GW (2010). Analyzing Qualitative Data: Systematic Approaches. Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Bradbury-Jones C, Breckenridge J, Clark MT, Herber OR, Wagstaff C, & Taylor J (2017). The state of qualitative research in health and social science literature: a focused mapping review and synthesis. International Journal of Social Research Methodology, 20(6), 627–645. [ Google Scholar ]
- Charmaz K (2000). Grounded theory: Objectivist and constructivist methods In Denzin NK & Lincoln YS (Eds.), Handbook of qualitative research (2nd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Charmaz K (2006). Constructing Grounded Theory: A Practical Guide through Qualitative Analysis. London: Sage Publications Ltd. [ Google Scholar ]
- Corbin JM, & Strauss A (1990). Grounded theory research: Procedures, canons, and evaluative criteria. Qualitative Sociology, 13(1), 3–21. [ Google Scholar ]
- Creswell JW (1998). Qualitative Inquiry and Research Design: Choosing Among Five Traditions. Thousand Oaks, CA: SAGE Publications Inc. [ Google Scholar ]
- Creswell JW (2003). Research Design: Qualitative, Quantitative, and Mixed Methods Approaches (2nd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Creswell JW (2007). Qualitative Inquiry and Research Design: Choosing Among Five Approaches (2nd ed.). Thousand Oaks, CA: SAGE. [ Google Scholar ]
- Creswell JW, & Poth CN (2018). Qualitative Inquiry & Research Design: Choosing Among Five Approaches (4th ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Guba E (1978) Toward a methodology of naturalistic inquiry in educational evaluation CSE Monograph Series in Evaluation. Los Angeles: University of California, UCLA Graduate School of Education. [ Google Scholar ]
- Hennink M, Hutter I, & Bailey A (2011). Qualitative Research Methods. London: SAGE Publications Ltd. [ Google Scholar ]
- Kegler MC, Raskind IG, Comeau DL, Griffith DM, Cooper HLF, & Shelton RC (2018). Study Design and Use of Inquiry Frameworks in Qualitative Research Published in Health Education & Behavior. Health Education and Behavior, 0(0), 1090198118795018. doi: 10.1177/1090198118795018 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
- Krueger R (1994). Focus groups: A practical guide for applied research (2nd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Krueger R (1998) Analyzing & Reporting Focus Group Results Focus Group Kit 6 Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Krueger R, & Casey MA (2000). Focus groups: A practical guide for applied research (3rd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Levitt HM, Bamberg M, Creswell JW, Frost DM, Josselson R, & Suarez-Orozco C (2018). Journal article reporting standards for qualitative primary, qualitative meta-analytic, and mixed methods research in psychology: The APA Publications and Communications Board task force report. American Psychologist, 73(1), 26–46. [ DOI ] [ PubMed ] [ Google Scholar ]
- Lincoln YS, & Guba E (1985). Naturalistic Inquiry. Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Maxwell J (1996). Qualitative Research Design: An Interactive Approach. Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Miles MB, & Huberman AM (1984). Qualitative Data Analysis: A Sourcebook of New Methods. Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Miles MB, & Huberman AM (1994). Qualitative Data Analysis: An Expanded Sourcebook (2nd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Miles MB, Huberman AM, & Saldana J (2014). Qualitative Data Analysis: A Methods Sourcebook (3rd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- O’Brien BC, Harris IB, Beckman TJ, Reed DA, & Cook DA (2014). Standards for reporting qualitative research: a synthesis of recommendations. Acad Med, 89(9), 1245–1251. [ DOI ] [ PubMed ] [ Google Scholar ]
- Patton MQ (1987). How to Use Qualitative Methods in Evaluation. Newbury Park, CA: SAGE Publications, Inc. [ Google Scholar ]
- Patton MQ (1990). Qualitative Evaluation and Research Methods (2nd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Patton MQ (1999). Enhancing the quality and credibility of qualitative analysis. Health Services Research, 34(5 Pt 2), 1189–1208. [ PMC free article ] [ PubMed ] [ Google Scholar ]
- Patton MQ (2002). Qualitative Research & Evaluation Methods (3rd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Patton MQ (2015). Qualitative Research & Evaluation Methods (4th ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Ryan GW, & Bernard HR (2000). Data management and analysis methods In Denzin NK & Lincoln YS (Eds.), Handbook of qualitative research (2nd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Ryan GW, & Bernard HR (2003). Techniques to Identify Themes. Field Methods, 15(1), 85–109. [ Google Scholar ]
- Saldana J (2016). The Coding Manual for Qualitative Researchers (3rd ed.). London: SAGE Publications Ltd. [ Google Scholar ]
- Saldana J, & Omasta M (2018). Qualitative Research: Analyzing Life. Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Spencer L, Ritchie J, Ormston R, O’Connor W, & Barnard M (2014). Analysis: Principles and Processes In Ritchie J, Lewis J, Nichols CM, & Ormston R (Eds.), Qualitative Research Practice: A Guide for Social Science Students & Researchers (2nd ed.). London: SAGE Publications Ltd. [ Google Scholar ]
- Strauss A, & Corbin JM (1994). Grounded theory methodology In Denzin NK & Lincoln YS (Eds.), Handbook of Qualitative Research. Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Strauss A, & Corbin JM (1998). Basics of Qualitative Research (2nd ed.). Thousand Oaks, CA: SAGE Publications, Inc. [ Google Scholar ]
- Tong A, Sainsbury P, & Craig J (2007). Consolidated criteria for reporting qualitative research (COREQ): a 32-item checklist for interviews and focus groups. International Journal for Quality in Health Care, 19(6), 349–357. [ DOI ] [ PubMed ] [ Google Scholar ]
- View on publisher site
- PDF (196.9 KB)
- Collections
Similar articles
Cited by other articles, links to ncbi databases.
- Download .nbib .nbib
- Format: AMA APA MLA NLM
Add to Collections
- - Google Chrome
Intended for healthcare professionals
- Access provided by Google Indexer
- My email alerts
- BMA member login
- Username * Password * Forgot your log in details? Need to activate BMA Member Log In Log in via OpenAthens Log in via your institution
Search form
- Advanced search
- Search responses
- Search blogs
- Practical thematic...
Practical thematic analysis: a guide for multidisciplinary health services research teams engaging in qualitative analysis
- Related content
- Peer review
- on behalf of the Coproduction Laboratory
- 1 Dartmouth Health, Lebanon, NH, USA
- 2 Dartmouth Institute for Health Policy and Clinical Practice, Geisel School of Medicine at Dartmouth College, Lebanon, NH, USA
- 3 Center for Primary Care and Public Health (Unisanté), Lausanne, Switzerland
- 4 Jönköping Academy for Improvement of Health and Welfare, School of Health and Welfare, Jönköping University, Jönköping, Sweden
- 5 Highland Park, NJ, USA
- 6 Division of Public Health Sciences, Department of Surgery, Washington University School of Medicine, St Louis, MO, USA
- Correspondence to: C H Saunders catherine.hylas.saunders{at}dartmouth.edu
- Accepted 26 April 2023
Qualitative research methods explore and provide deep contextual understanding of real world issues, including people’s beliefs, perspectives, and experiences. Whether through analysis of interviews, focus groups, structured observation, or multimedia data, qualitative methods offer unique insights in applied health services research that other approaches cannot deliver. However, many clinicians and researchers hesitate to use these methods, or might not use them effectively, which can leave relevant areas of inquiry inadequately explored. Thematic analysis is one of the most common and flexible methods to examine qualitative data collected in health services research. This article offers practical thematic analysis as a step-by-step approach to qualitative analysis for health services researchers, with a focus on accessibility for patients, care partners, clinicians, and others new to thematic analysis. Along with detailed instructions covering three steps of reading, coding, and theming, the article includes additional novel and practical guidance on how to draft effective codes, conduct a thematic analysis session, and develop meaningful themes. This approach aims to improve consistency and rigor in thematic analysis, while also making this method more accessible for multidisciplinary research teams.
Through qualitative methods, researchers can provide deep contextual understanding of real world issues, and generate new knowledge to inform hypotheses, theories, research, and clinical care. Approaches to data collection are varied, including interviews, focus groups, structured observation, and analysis of multimedia data, with qualitative research questions aimed at understanding the how and why of human experience. 1 2 Qualitative methods produce unique insights in applied health services research that other approaches cannot deliver. In particular, researchers acknowledge that thematic analysis is a flexible and powerful method of systematically generating robust qualitative research findings by identifying, analysing, and reporting patterns (themes) within data. 3 4 5 6 Although qualitative methods are increasingly valued for answering clinical research questions, many researchers are unsure how to apply them or consider them too time consuming to be useful in responding to practical challenges 7 or pressing situations such as public health emergencies. 8 Consequently, researchers might hesitate to use them, or use them improperly. 9 10 11
Although much has been written about how to perform thematic analysis, practical guidance for non-specialists is sparse. 3 5 6 12 13 In the multidisciplinary field of health services research, qualitative data analysis can confound experienced researchers and novices alike, which can stoke concerns about rigor, particularly for those more familiar with quantitative approaches. 14 Since qualitative methods are an area of specialisation, support from experts is beneficial. However, because non-specialist perspectives can enhance data interpretation and enrich findings, there is a case for making thematic analysis easier, more rapid, and more efficient, 8 particularly for patients, care partners, clinicians, and other stakeholders. A practical guide to thematic analysis might encourage those on the ground to use these methods in their work, unearthing insights that would otherwise remain undiscovered.
Given the need for more accessible qualitative analysis approaches, we present a simple, rigorous, and efficient three step guide for practical thematic analysis. We include new guidance on the mechanics of thematic analysis, including developing codes, constructing meaningful themes, and hosting a thematic analysis session. We also discuss common pitfalls in thematic analysis and how to avoid them.
Summary points
Qualitative methods are increasingly valued in applied health services research, but multidisciplinary research teams often lack accessible step-by-step guidance and might struggle to use these approaches
A newly developed approach, practical thematic analysis, uses three simple steps: reading, coding, and theming
Based on Braun and Clarke’s reflexive thematic analysis, our streamlined yet rigorous approach is designed for multidisciplinary health services research teams, including patients, care partners, and clinicians
This article also provides companion materials including a slide presentation for teaching practical thematic analysis to research teams, a sample thematic analysis session agenda, a theme coproduction template for use during the session, and guidance on using standardised reporting criteria for qualitative research
In their seminal work, Braun and Clarke developed a six phase approach to reflexive thematic analysis. 4 12 We built on their method to develop practical thematic analysis ( box 1 , fig 1 ), which is a simplified and instructive approach that retains the substantive elements of their six phases. Braun and Clarke’s phase 1 (familiarising yourself with the dataset) is represented in our first step of reading. Phase 2 (coding) remains as our second step of coding. Phases 3 (generating initial themes), 4 (developing and reviewing themes), and 5 (refining, defining, and naming themes) are represented in our third step of theming. Phase 6 (writing up) also occurs during this third step of theming, but after a thematic analysis session. 4 12
Key features and applications of practical thematic analysis
Step 1: reading.
All manuscript authors read the data
All manuscript authors write summary memos
Step 2: Coding
Coders perform both data management and early data analysis
Codes are complete thoughts or sentences, not categories
Step 3: Theming
Researchers host a thematic analysis session and share different perspectives
Themes are complete thoughts or sentences, not categories
Applications
For use by practicing clinicians, patients and care partners, students, interdisciplinary teams, and those new to qualitative research
When important insights from healthcare professionals are inaccessible because they do not have qualitative methods training
When time and resources are limited
Steps in practical thematic analysis
- Download figure
- Open in new tab
- Download powerpoint
We present linear steps, but as qualitative research is usually iterative, so too is thematic analysis. 15 Qualitative researchers circle back to earlier work to check whether their interpretations still make sense in the light of additional insights, adapting as necessary. While we focus here on the practical application of thematic analysis in health services research, we recognise our approach exists in the context of the broader literature on thematic analysis and the theoretical underpinnings of qualitative methods as a whole. For a more detailed discussion of these theoretical points, as well as other methods widely used in health services research, we recommend reviewing the sources outlined in supplemental material 1. A strong and nuanced understanding of the context and underlying principles of thematic analysis will allow for higher quality research. 16
Practical thematic analysis is a highly flexible approach that can draw out valuable findings and generate new hypotheses, including in cases with a lack of previous research to build on. The approach can also be used with a variety of data, such as transcripts from interviews or focus groups, patient encounter transcripts, professional publications, observational field notes, and online activity logs. Importantly, successful practical thematic analysis is predicated on having high quality data collected with rigorous methods. We do not describe qualitative research design or data collection here. 11 17
In supplemental material 1, we summarise the foundational methods, concepts, and terminology in qualitative research. Along with our guide below, we include a companion slide presentation for teaching practical thematic analysis to research teams in supplemental material 2. We provide a theme coproduction template for teams to use during thematic analysis sessions in supplemental material 3. Our method aligns with the major qualitative reporting frameworks, including the Consolidated Criteria for Reporting Qualitative Research (COREQ). 18 We indicate the corresponding step in practical thematic analysis for each COREQ item in supplemental material 4.
Familiarisation and memoing
We encourage all manuscript authors to review the full dataset (eg, interview transcripts) to familiarise themselves with it. This task is most critical for those who will later be engaged in the coding and theming steps. Although time consuming, it is the best way to involve team members in the intellectual work of data interpretation, so that they can contribute to the analysis and contextualise the results. If this task is not feasible given time limitations or large quantities of data, the data can be divided across team members. In this case, each piece of data should be read by at least two individuals who ideally represent different professional roles or perspectives.
We recommend that researchers reflect on the data and independently write memos, defined as brief notes on thoughts and questions that arise during reading, and a summary of their impressions of the dataset. 2 19 Memoing is an opportunity to gain insights from varying perspectives, particularly from patients, care partners, clinicians, and others. It also gives researchers the opportunity to begin to scope which elements of and concepts in the dataset are relevant to the research question.
Data saturation
The concept of data saturation ( box 2 ) is a foundation of qualitative research. It is defined as the point in analysis at which new data tend to be redundant of data already collected. 21 Qualitative researchers are expected to report their approach to data saturation. 18 Because thematic analysis is iterative, the team should discuss saturation throughout the entire process, beginning with data collection and continuing through all steps of the analysis. 22 During step 1 (reading), team members might discuss data saturation in the context of summary memos. Conversations about saturation continue during step 2 (coding), with confirmation that saturation has been achieved during step 3 (theming). As a rule of thumb, researchers can often achieve saturation in 9-17 interviews or 4-8 focus groups, but this will vary depending on the specific characteristics of the study. 23
Data saturation in context
Braun and Clarke discourage the use of data saturation to determine sample size (eg, number of interviews), because it assumes that there is an objective truth to be captured in the data (sometimes known as a positivist perspective). 20 Qualitative researchers often try to avoid positivist approaches, arguing that there is no one true way of seeing the world, and will instead aim to gather multiple perspectives. 5 Although this theoretical debate with qualitative methods is important, we recognise that a priori estimates of saturation are often needed, particularly for investigators newer to qualitative research who might want a more pragmatic and applied approach. In addition, saturation based, sample size estimation can be particularly helpful in grant proposals. However, researchers should still follow a priori sample size estimation with a discussion to confirm saturation has been achieved.
Definition of coding
We describe codes as labels for concepts in the data that are directly relevant to the study objective. Historically, the purpose of coding was to distil the large amount of data collected into conceptually similar buckets so that researchers could review it in aggregate and identify key themes. 5 24 We advocate for a more analytical approach than is typical with thematic analysis. With our method, coding is both the foundation for and the beginning of thematic analysis—that is, early data analysis, management, and reduction occur simultaneously rather than as different steps. This approach moves the team more efficiently towards being able to describe themes.
Building the coding team
Coders are the research team members who directly assign codes to the data, reading all material and systematically labelling relevant data with appropriate codes. Ideally, at least two researchers would code every discrete data document, such as one interview transcript. 25 If this task is not possible, individual coders can each code a subset of the data that is carefully selected for key characteristics (sometimes known as purposive selection). 26 When using this approach, we recommend that at least 10% of data be coded by two or more coders to ensure consistency in codebook application. We also recommend coding teams of no more than four to five people, for practical reasons concerning maintaining consistency.
Clinicians, patients, and care partners bring unique perspectives to coding and enrich the analytical process. 27 Therefore, we recommend choosing coders with a mix of relevant experiences so that they can challenge and contextualise each other’s interpretations based on their own perspectives and opinions ( box 3 ). We recommend including both coders who collected the data and those who are naive to it, if possible, given their different perspectives. We also recommend all coders review the summary memos from the reading step so that key concepts identified by those not involved in coding can be integrated into the analytical process. In practice, this review means coding the memos themselves and discussing them during the code development process. This approach ensures that the team considers a diversity of perspectives.
Coding teams in context
The recommendation to use multiple coders is a departure from Braun and Clarke. 28 29 When the views, experiences, and training of each coder (sometimes known as positionality) 30 are carefully considered, having multiple coders can enhance interpretation and enrich findings. When these perspectives are combined in a team setting, researchers can create shared meaning from the data. Along with the practical consideration of distributing the workload, 31 inclusion of these multiple perspectives increases the overall quality of the analysis by mitigating the impact of any one coder’s perspective. 30
Coding tools
Qualitative analysis software facilitates coding and managing large datasets but does not perform the analytical work. The researchers must perform the analysis themselves. Most programs support queries and collaborative coding by multiple users. 32 Important factors to consider when choosing software can include accessibility, cost, interoperability, the look and feel of code reports, and the ease of colour coding and merging codes. Coders can also use low tech solutions, including highlighters, word processors, or spreadsheets.
Drafting effective codes
To draft effective codes, we recommend that the coders review each document line by line. 33 As they progress, they can assign codes to segments of data representing passages of interest. 34 Coders can also assign multiple codes to the same passage. Consensus among coders on what constitutes a minimum or maximum amount of text for assigning a code is helpful. As a general rule, meaningful segments of text for coding are shorter than one paragraph, but longer than a few words. Coders should keep the study objective in mind when determining which data are relevant ( box 4 ).
Code types in context
Similar to Braun and Clarke’s approach, practical thematic analysis does not specify whether codes are based on what is evident from the data (sometimes known as semantic) or whether they are based on what can be inferred at a deeper level from the data (sometimes known as latent). 4 12 35 It also does not specify whether they are derived from the data (sometimes known as inductive) or determined ahead of time (sometimes known as deductive). 11 35 Instead, it should be noted that health services researchers conducting qualitative studies often adopt all these approaches to coding (sometimes known as hybrid analysis). 3
In practical thematic analysis, codes should be more descriptive than general categorical labels that simply group data with shared characteristics. At a minimum, codes should form a complete (or full) thought. An easy way to conceptualise full thought codes is as complete sentences with subjects and verbs ( table 1 ), although full sentence coding is not always necessary. With full thought codes, researchers think about the data more deeply and capture this insight in the codes. This coding facilitates the entire analytical process and is especially valuable when moving from codes to broader themes. Experienced qualitative researchers often intuitively use full thought or sentence codes, but this practice has not been explicitly articulated as a path to higher quality coding elsewhere in the literature. 6
Example transcript with codes used in practical thematic analysis 36
- View inline
Depending on the nature of the data, codes might either fall into flat categories or be arranged hierarchically. Flat categories are most common when the data deal with topics on the same conceptual level. In other words, one topic is not a subset of another topic. By contrast, hierarchical codes are more appropriate for concepts that naturally fall above or below each other. Hierarchical coding can also be a useful form of data management and might be necessary when working with a large or complex dataset. 5 Codes grouped into these categories can also make it easier to naturally transition into generating themes from the initial codes. 5 These decisions between flat versus hierarchical coding are part of the work of the coding team. In both cases, coders should ensure that their code structures are guided by their research questions.
Developing the codebook
A codebook is a shared document that lists code labels and comprehensive descriptions for each code, as well as examples observed within the data. Good code descriptions are precise and specific so that coders can consistently assign the same codes to relevant data or articulate why another coder would do so. Codebook development is iterative and involves input from the entire coding team. However, as those closest to the data, coders must resist undue influence, real or perceived, from other team members with conflicting opinions—it is important to mitigate the risk that more senior researchers, like principal investigators, exert undue influence on the coders’ perspectives.
In practical thematic analysis, coders begin codebook development by independently coding a small portion of the data, such as two to three transcripts or other units of analysis. Coders then individually produce their initial codebooks. This task will require them to reflect on, organise, and clarify codes. The coders then meet to reconcile the draft codebooks, which can often be difficult, as some coders tend to lump several concepts together while others will split them into more specific codes. Discussing disagreements and negotiating consensus are necessary parts of early data analysis. Once the codebook is relatively stable, we recommend soliciting input on the codes from all manuscript authors. Yet, coders must ultimately be empowered to finalise the details so that they are comfortable working with the codebook across a large quantity of data.
Assigning codes to the data
After developing the codebook, coders will use it to assign codes to the remaining data. While the codebook’s overall structure should remain constant, coders might continue to add codes corresponding to any new concepts observed in the data. If new codes are added, coders should review the data they have already coded and determine whether the new codes apply. Qualitative data analysis software can be useful for editing or merging codes.
We recommend that coders periodically compare their code occurrences ( box 5 ), with more frequent check-ins if substantial disagreements occur. In the event of large discrepancies in the codes assigned, coders should revise the codebook to ensure that code descriptions are sufficiently clear and comprehensive to support coding alignment going forward. Because coding is an iterative process, the team can adjust the codebook as needed. 5 28 29
Quantitative coding in context
Researchers should generally avoid reporting code counts in thematic analysis. However, counts can be a useful proxy in maintaining alignment between coders on key concepts. 26 In practice, therefore, researchers should make sure that all coders working on the same piece of data assign the same codes with a similar pattern and that their memoing and overall assessment of the data are aligned. 37 However, the frequency of a code alone is not an indicator of its importance. It is more important that coders agree on the most salient points in the data; reviewing and discussing summary memos can be helpful here. 5
Researchers might disagree on whether or not to calculate and report inter-rater reliability. We note that quantitative tests for agreement, such as kappa statistics or intraclass correlation coefficients, can be distracting and might not provide meaningful results in qualitative analyses. Similarly, Braun and Clarke argue that expecting perfect alignment on coding is inconsistent with the goal of co-constructing meaning. 28 29 Overall consensus on codes’ salience and contributions to themes is the most important factor.
Definition of themes
Themes are meta-constructs that rise above codes and unite the dataset ( box 6 , fig 2 ). They should be clearly evident, repeated throughout the dataset, and relevant to the research questions. 38 While codes are often explicit descriptions of the content in the dataset, themes are usually more conceptual and knit the codes together. 39 Some researchers hypothesise that theme development is loosely described in the literature because qualitative researchers simply intuit themes during the analytical process. 39 In practical thematic analysis, we offer a concrete process that should make developing meaningful themes straightforward.
Themes in context
According to Braun and Clarke, a theme “captures something important about the data in relation to the research question and represents some level of patterned response or meaning within the data set.” 4 Similarly, Braun and Clarke advise against themes as domain summaries. While different approaches can draw out themes from codes, the process begins by identifying patterns. 28 35 Like Braun and Clarke and others, we recommend that researchers consider the salience of certain themes, their prevalence in the dataset, and their keyness (ie, how relevant the themes are to the overarching research questions). 4 12 34
Use of themes in practical thematic analysis
Constructing meaningful themes
After coding all the data, each coder should independently reflect on the team’s summary memos (step 1), the codebook (step 2), and the coded data itself to develop draft themes (step 3). It can be illuminating for coders to review all excerpts associated with each code, so that they derive themes directly from the data. Researchers should remain focused on the research question during this step, so that themes have a clear relation with the overall project aim. Use of qualitative analysis software will make it easy to view each segment of data tagged with each code. Themes might neatly correspond to groups of codes. Or—more likely—they will unite codes and data in unexpected ways. A whiteboard or presentation slides might be helpful to organise, craft, and revise themes. We also provide a template for coproducing themes (supplemental material 3). As with codebook justification, team members will ideally produce individual drafts of the themes that they have identified in the data. They can then discuss these with the group and reach alignment or consensus on the final themes.
The team should ensure that all themes are salient, meaning that they are: supported by the data, relevant to the study objectives, and important. Similar to codes, themes are framed as complete thoughts or sentences, not categories. While codes and themes might appear to be similar to each other, the key distinction is that the themes represent a broader concept. Table 2 shows examples of codes and their corresponding themes from a previously published project that used practical thematic analysis. 36 Identifying three to four key themes that comprise a broader overarching theme is a useful approach. Themes can also have subthemes, if appropriate. 40 41 42 43 44
Example codes with themes in practical thematic analysis 36
Thematic analysis session
After each coder has independently produced draft themes, a carefully selected subset of the manuscript team meets for a thematic analysis session ( table 3 ). The purpose of this session is to discuss and reach alignment or consensus on the final themes. We recommend a session of three to five hours, either in-person or virtually.
Example agenda of thematic analysis session
The composition of the thematic analysis session team is important, as each person’s perspectives will shape the results. This group is usually a small subset of the broader research team, with three to seven individuals. We recommend that primary and senior authors work together to include people with diverse experiences related to the research topic. They should aim for a range of personalities and professional identities, particularly those of clinicians, trainees, patients, and care partners. At a minimum, all coders and primary and senior authors should participate in the thematic analysis session.
The session begins with each coder presenting their draft themes with supporting quotes from the data. 5 Through respectful and collaborative deliberation, the group will develop a shared set of final themes.
One team member facilitates the session. A firm, confident, and consistent facilitation style with good listening skills is critical. For practical reasons, this person is not usually one of the primary coders. Hierarchies in teams cannot be entirely flattened, but acknowledging them and appointing an external facilitator can reduce their impact. The facilitator can ensure that all voices are heard. For example, they might ask for perspectives from patient partners or more junior researchers, and follow up on comments from senior researchers to say, “We have heard your perspective and it is important; we want to make sure all perspectives in the room are equally considered.” Or, “I hear [senior person] is offering [x] idea, I’d like to hear other perspectives in the room.” The role of the facilitator is critical in the thematic analysis session. The facilitator might also privately discuss with more senior researchers, such as principal investigators and senior authors, the importance of being aware of their influence over others and respecting and eliciting the perspectives of more junior researchers, such as patients, care partners, and students.
To our knowledge, this discrete thematic analysis session is a novel contribution of practical thematic analysis. It helps efficiently incorporate diverse perspectives using the session agenda and theme coproduction template (supplemental material 3) and makes the process of constructing themes transparent to the entire research team.
Writing the report
We recommend beginning the results narrative with a summary of all relevant themes emerging from the analysis, followed by a subheading for each theme. Each subsection begins with a brief description of the theme and is illustrated with relevant quotes, which are contextualised and explained. The write-up should not simply be a list, but should contain meaningful analysis and insight from the researchers, including descriptions of how different stakeholders might have experienced a particular situation differently or unexpectedly.
In addition to weaving quotes into the results narrative, quotes can be presented in a table. This strategy is a particularly helpful when submitting to clinical journals with tight word count limitations. Quote tables might also be effective in illustrating areas of agreement and disagreement across stakeholder groups, with columns representing different groups and rows representing each theme or subtheme. Quotes should include an anonymous label for each participant and any relevant characteristics, such as role or gender. The aim is to produce rich descriptions. 5 We recommend against repeating quotations across multiple themes in the report, so as to avoid confusion. The template for coproducing themes (supplemental material 3) allows documentation of quotes supporting each theme, which might also be useful during report writing.
Visual illustrations such as a thematic map or figure of the findings can help communicate themes efficiently. 4 36 42 44 If a figure is not possible, a simple list can suffice. 36 Both must clearly present the main themes with subthemes. Thematic figures can facilitate confirmation that the researchers’ interpretations reflect the study populations’ perspectives (sometimes known as member checking), because authors can invite discussions about the figure and descriptions of findings and supporting quotes. 46 This process can enhance the validity of the results. 46
In supplemental material 4, we provide additional guidance on reporting thematic analysis consistent with COREQ. 18 Commonly used in health services research, COREQ outlines a standardised list of items to be included in qualitative research reports ( box 7 ).
Reporting in context
We note that use of COREQ or any other reporting guidelines does not in itself produce high quality work and should not be used as a substitute for general methodological rigor. Rather, researchers must consider rigor throughout the entire research process. As the issue of how to conceptualise and achieve rigorous qualitative research continues to be debated, 47 48 we encourage researchers to explicitly discuss how they have looked at methodological rigor in their reports. Specifically, we point researchers to Braun and Clarke’s 2021 tool for evaluating thematic analysis manuscripts for publication (“Twenty questions to guide assessment of TA [thematic analysis] research quality”). 16
Avoiding common pitfalls
Awareness of common mistakes can help researchers avoid improper use of qualitative methods. Improper use can, for example, prevent researchers from developing meaningful themes and can risk drawing inappropriate conclusions from the data. Braun and Clarke also warn of poor quality in qualitative research, noting that “coherence and integrity of published research does not always hold.” 16
Weak themes
An important distinction between high and low quality themes is that high quality themes are descriptive and complete thoughts. As such, they often contain subjects and verbs, and can be expressed as full sentences ( table 2 ). Themes that are simply descriptive categories or topics could fail to impart meaningful knowledge beyond categorisation. 16 49 50
Researchers will often move from coding directly to writing up themes, without performing the work of theming or hosting a thematic analysis session. Skipping concerted theming often results in themes that look more like categories than unifying threads across the data.
Unfocused analysis
Because data collection for qualitative research is often semi-structured (eg, interviews, focus groups), not all data will be directly relevant to the research question at hand. To avoid unfocused analysis and a correspondingly unfocused manuscript, we recommend that all team members keep the research objective in front of them at every stage, from reading to coding to theming. During the thematic analysis session, we recommend that the research question be written on a whiteboard so that all team members can refer back to it, and so that the facilitator can ensure that conversations about themes occur in the context of this question. Consistently focusing on the research question can help to ensure that the final report directly answers it, as opposed to the many other interesting insights that might emerge during the qualitative research process. Such insights can be picked up in a secondary analysis if desired.
Inappropriate quantification
Presenting findings quantitatively (eg, “We found 18 instances of participants mentioning safety concerns about the vaccines”) is generally undesirable in practical thematic analysis reporting. 51 Descriptive terms are more appropriate (eg, “participants had substantial concerns about the vaccines,” or “several participants were concerned about this”). This descriptive presentation is critical because qualitative data might not be consistently elicited across participants, meaning that some individuals might share certain information while others do not, simply based on how conversations evolve. Additionally, qualitative research does not aim to draw inferences outside its specific sample. Emphasising numbers in thematic analysis can lead to readers incorrectly generalising the findings. Although peer reviewers unfamiliar with thematic analysis often request this type of quantification, practitioners of practical thematic analysis can confidently defend their decision to avoid it. If quantification is methodologically important, we recommend simultaneously conducting a survey or incorporating standardised interview techniques into the interview guide. 11
Neglecting group dynamics
Researchers should concertedly consider group dynamics in the research team. Particular attention should be paid to power relations and the personality of team members, which can include aspects such as who most often speaks, who defines concepts, and who resolves disagreements that might arise within the group. 52
The perspectives of patient and care partners are particularly important to cultivate. Ideally, patient partners are meaningfully embedded in studies from start to finish, not just for practical thematic analysis. 53 Meaningful engagement can build trust, which makes it easier for patient partners to ask questions, request clarification, and share their perspectives. Professional team members should actively encourage patient partners by emphasising that their expertise is critically important and valued. Noting when a patient partner might be best positioned to offer their perspective can be particularly powerful.
Insufficient time allocation
Researchers must allocate enough time to complete thematic analysis. Working with qualitative data takes time, especially because it is often not a linear process. As the strength of thematic analysis lies in its ability to make use of the rich details and complexities of the data, we recommend careful planning for the time required to read and code each document.
Estimating the necessary time can be challenging. For step 1 (reading), researchers can roughly calculate the time required based on the time needed to read and reflect on one piece of data. For step 2 (coding), the total amount of time needed can be extrapolated from the time needed to code one document during codebook development. We also recommend three to five hours for the thematic analysis session itself, although coders will need to independently develop their draft themes beforehand. Although the time required for practical thematic analysis is variable, teams should be able to estimate their own required effort with these guidelines.
Practical thematic analysis builds on the foundational work of Braun and Clarke. 4 16 We have reframed their six phase process into three condensed steps of reading, coding, and theming. While we have maintained important elements of Braun and Clarke’s reflexive thematic analysis, we believe that practical thematic analysis is conceptually simpler and easier to teach to less experienced researchers and non-researcher stakeholders. For teams with different levels of familiarity with qualitative methods, this approach presents a clear roadmap to the reading, coding, and theming of qualitative data. Our practical thematic analysis approach promotes efficient learning by doing—experiential learning. 12 29 Practical thematic analysis avoids the risk of relying on complex descriptions of methods and theory and places more emphasis on obtaining meaningful insights from those close to real world clinical environments. Although practical thematic analysis can be used to perform intensive theory based analyses, it lends itself more readily to accelerated, pragmatic approaches.
Strengths and limitations
Our approach is designed to smooth the qualitative analysis process and yield high quality themes. Yet, researchers should note that poorly performed analyses will still produce low quality results. Practical thematic analysis is a qualitative analytical approach; it does not look at study design, data collection, or other important elements of qualitative research. It also might not be the right choice for every qualitative research project. We recommend it for applied health services research questions, where diverse perspectives and simplicity might be valuable.
We also urge researchers to improve internal validity through triangulation methods, such as member checking (supplemental material 1). 46 Member checking could include soliciting input on high level themes, theme definitions, and quotations from participants. This approach might increase rigor.
Implications
We hope that by providing clear and simple instructions for practical thematic analysis, a broader range of researchers will be more inclined to use these methods. Increased transparency and familiarity with qualitative approaches can enhance researchers’ ability to both interpret qualitative studies and offer up new findings themselves. In addition, it can have usefulness in training and reporting. A major strength of this approach is to facilitate meaningful inclusion of patient and care partner perspectives, because their lived experiences can be particularly valuable in data interpretation and the resulting findings. 11 30 As clinicians are especially pressed for time, they might also appreciate a practical set of instructions that can be immediately used to leverage their insights and access to patients and clinical settings, and increase the impact of qualitative research through timely results. 8
Practical thematic analysis is a simplified approach to performing thematic analysis in health services research, a field where the experiences of patients, care partners, and clinicians are of inherent interest. We hope that it will be accessible to those individuals new to qualitative methods, including patients, care partners, clinicians, and other health services researchers. We intend to empower multidisciplinary research teams to explore unanswered questions and make new, important, and rigorous contributions to our understanding of important clinical and health systems research.
Acknowledgments
All members of the Coproduction Laboratory provided input that shaped this manuscript during laboratory meetings. We acknowledge advice from Elizabeth Carpenter-Song, an expert in qualitative methods.
Coproduction Laboratory group contributors: Stephanie C Acquilano ( http://orcid.org/0000-0002-1215-5531 ), Julie Doherty ( http://orcid.org/0000-0002-5279-6536 ), Rachel C Forcino ( http://orcid.org/0000-0001-9938-4830 ), Tina Foster ( http://orcid.org/0000-0001-6239-4031 ), Megan Holthoff, Christopher R Jacobs ( http://orcid.org/0000-0001-5324-8657 ), Lisa C Johnson ( http://orcid.org/0000-0001-7448-4931 ), Elaine T Kiriakopoulos, Kathryn Kirkland ( http://orcid.org/0000-0002-9851-926X ), Meredith A MacMartin ( http://orcid.org/0000-0002-6614-6091 ), Emily A Morgan, Eugene Nelson, Elizabeth O’Donnell, Brant Oliver ( http://orcid.org/0000-0002-7399-622X ), Danielle Schubbe ( http://orcid.org/0000-0002-9858-1805 ), Gabrielle Stevens ( http://orcid.org/0000-0001-9001-178X ), Rachael P Thomeer ( http://orcid.org/0000-0002-5974-3840 ).
Contributors: Practical thematic analysis, an approach designed for multidisciplinary health services teams new to qualitative research, was based on CHS’s experiences teaching thematic analysis to clinical teams and students. We have drawn heavily from qualitative methods literature. CHS is the guarantor of the article. CHS, AS, CvP, AMK, JRK, and JAP contributed to drafting the manuscript. AS, JG, CMM, JAP, and RWY provided feedback on their experiences using practical thematic analysis. CvP, LCL, SLB, AVC, GE, and JKL advised on qualitative methods in health services research, given extensive experience. All authors meaningfully edited the manuscript content, including AVC and RKS. The corresponding author attests that all listed authors meet authorship criteria and that no others meeting the criteria have been omitted.
Funding: This manuscript did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.
Competing interests: All authors have completed the ICMJE uniform disclosure form at https://www.icmje.org/disclosure-of-interest/ and declare: no support from any organisation for the submitted work; no financial relationships with any organisations that might have an interest in the submitted work in the previous three years; no other relationships or activities that could appear to have influenced the submitted work.
Provenance and peer review: Not commissioned; externally peer reviewed.
- Ziebland S ,
- ↵ A Hybrid Approach to Thematic Analysis in Qualitative Research: Using a Practical Example. 2018. https://methods.sagepub.com/case/hybrid-approach-thematic-analysis-qualitative-research-a-practical-example .
- Maguire M ,
- Vindrola-Padros C ,
- Vindrola-Padros B
- ↵ Vindrola-Padros C. Rapid Ethnographies: A Practical Guide . Cambridge University Press 2021. https://play.google.com/store/books/details?id=n80HEAAAQBAJ
- Schroter S ,
- Merino JG ,
- Barbeau A ,
- ↵ Padgett DK. Qualitative and Mixed Methods in Public Health . SAGE Publications 2011. https://play.google.com/store/books/details?id=LcYgAQAAQBAJ
- Scharp KM ,
- Korstjens I
- Barnett-Page E ,
- ↵ Guest G, Namey EE, Mitchell ML. Collecting Qualitative Data: A Field Manual for Applied Research . SAGE 2013. https://play.google.com/store/books/details?id=-3rmWYKtloC
- Sainsbury P ,
- Emerson RM ,
- Saunders B ,
- Kingstone T ,
- Hennink MM ,
- Kaiser BN ,
- Hennink M ,
- O’Connor C ,
- ↵ Yen RW, Schubbe D, Walling L, et al. Patient engagement in the What Matters Most trial: experiences and future implications for research. Poster presented at International Shared Decision Making conference, Quebec City, Canada. July 2019.
- ↵ Got questions about Thematic Analysis? We have prepared some answers to common ones. https://www.thematicanalysis.net/faqs/ (accessed 9 Nov 2022).
- ↵ Braun V, Clarke V. Thematic Analysis. SAGE Publications. 2022. https://uk.sagepub.com/en-gb/eur/thematic-analysis/book248481 .
- Kalpokas N ,
- Radivojevic I
- Campbell KA ,
- Durepos P ,
- ↵ Understanding Thematic Analysis. https://www.thematicanalysis.net/understanding-ta/ .
- Saunders CH ,
- Stevens G ,
- CONFIDENT Study Long-Term Care Partners
- MacQueen K ,
- Vaismoradi M ,
- Turunen H ,
- Schott SL ,
- Berkowitz J ,
- Carpenter-Song EA ,
- Goldwag JL ,
- Durand MA ,
- Goldwag J ,
- Saunders C ,
- Mishra MK ,
- Rodriguez HP ,
- Shortell SM ,
- Verdinelli S ,
- Scagnoli NI
- Campbell C ,
- Sparkes AC ,
- McGannon KR
- Sandelowski M ,
- Connelly LM ,
- O’Malley AJ ,
IMAGES
VIDEO
COMMENTS
Whether or when to use qualitative methods in field epidemiology studies ultimately depends on the nature of the public health question to be answered. Qualitative approaches make sense when a study question about behavior patterns or program performance leads with why, why not , or how .
We discuss our practical experience with a team-based approach using flexible coding for qualitative data analysis in public health, illustrating how this process can be applied to address multiple research questions simultaneously or asynchronously.
In an era of public health grounded in evidence-based research and practice, rigorously conducting, documenting, and reporting qualitative analysis is critical for the generation of reliable and actionable knowledge.
In this article, we briefly review the role and use of qualitative methods in public health research and its significance for research, policy and practice. Historically, public health research has been largely dependent on quantitative research rooted in medical science.
We discuss our practical experience with a team-based approach using flexible coding for qualitative data analysis in public health, illustrating how this process can be applied to address...
This article offers practical thematic analysis as a step-by-step approach to qualitative analysis for health services researchers, with a focus on accessibility for patients, care partners, clinicians, and others new to thematic analysis.