Service update: Some parts of the Library’s website will be down for maintenance on August 11.
Secondary menu
- Log in to your Library account
- Hours and Maps
- Connect from Off Campus
- UC Berkeley Home
Search form
Research methods--quantitative, qualitative, and more: overview.
- Quantitative Research
- Qualitative Research
- Data Science Methods (Machine Learning, AI, Big Data)
- Text Mining and Computational Text Analysis
- Evidence Synthesis/Systematic Reviews
- Get Data, Get Help!
About Research Methods
This guide provides an overview of research methods, how to choose and use them, and supports and resources at UC Berkeley.
As Patten and Newhart note in the book Understanding Research Methods , "Research methods are the building blocks of the scientific enterprise. They are the "how" for building systematic knowledge. The accumulation of knowledge through research is by its nature a collective endeavor. Each well-designed study provides evidence that may support, amend, refute, or deepen the understanding of existing knowledge...Decisions are important throughout the practice of research and are designed to help researchers collect evidence that includes the full spectrum of the phenomenon under study, to maintain logical rules, and to mitigate or account for possible sources of bias. In many ways, learning research methods is learning how to see and make these decisions."
The choice of methods varies by discipline, by the kind of phenomenon being studied and the data being used to study it, by the technology available, and more. This guide is an introduction, but if you don't see what you need here, always contact your subject librarian, and/or take a look to see if there's a library research guide that will answer your question.
Suggestions for changes and additions to this guide are welcome!
START HERE: SAGE Research Methods
Without question, the most comprehensive resource available from the library is SAGE Research Methods. HERE IS THE ONLINE GUIDE to this one-stop shopping collection, and some helpful links are below:
- SAGE Research Methods
- Little Green Books (Quantitative Methods)
- Little Blue Books (Qualitative Methods)
- Dictionaries and Encyclopedias
- Case studies of real research projects
- Sample datasets for hands-on practice
- Streaming video--see methods come to life
- Methodspace- -a community for researchers
- SAGE Research Methods Course Mapping
Library Data Services at UC Berkeley
Library Data Services Program and Digital Scholarship Services
The LDSP offers a variety of services and tools ! From this link, check out pages for each of the following topics: discovering data, managing data, collecting data, GIS data, text data mining, publishing data, digital scholarship, open science, and the Research Data Management Program.
Be sure also to check out the visual guide to where to seek assistance on campus with any research question you may have!
Library GIS Services
Other Data Services at Berkeley
D-Lab Supports Berkeley faculty, staff, and graduate students with research in data intensive social science, including a wide range of training and workshop offerings Dryad Dryad is a simple self-service tool for researchers to use in publishing their datasets. It provides tools for the effective publication of and access to research data. Geospatial Innovation Facility (GIF) Provides leadership and training across a broad array of integrated mapping technologies on campu Research Data Management A UC Berkeley guide and consulting service for research data management issues
General Research Methods Resources
Here are some general resources for assistance:
- Assistance from ICPSR (must create an account to access): Getting Help with Data , and Resources for Students
- Wiley Stats Ref for background information on statistics topics
- Survey Documentation and Analysis (SDA) . Program for easy web-based analysis of survey data.
Consultants
- D-Lab/Data Science Discovery Consultants Request help with your research project from peer consultants.
- Research data (RDM) consulting Meet with RDM consultants before designing the data security, storage, and sharing aspects of your qualitative project.
- Statistics Department Consulting Services A service in which advanced graduate students, under faculty supervision, are available to consult during specified hours in the Fall and Spring semesters.
Related Resourcex
- IRB / CPHS Qualitative research projects with human subjects often require that you go through an ethics review.
- OURS (Office of Undergraduate Research and Scholarships) OURS supports undergraduates who want to embark on research projects and assistantships. In particular, check out their "Getting Started in Research" workshops
- Sponsored Projects Sponsored projects works with researchers applying for major external grants.
- Next: Quantitative Research >>
- Last Updated: Sep 6, 2024 8:59 PM
- URL: https://guides.lib.berkeley.edu/researchmethods
Strategies and Models
The choice of qualitative or quantitative approach to research has been traditionally guided by the subject discipline. However, this is changing, with many “applied” researchers taking a more holistic and integrated approach that combines the two traditions. This methodology reflects the multi-disciplinary nature of many contemporary research problems.
In fact, it is possible to define many different types of research strategy. The following list ( Business research methods / Alan Bryman & Emma Bell. 4th ed. Oxford : Oxford University Press, 2015 ) is neither exclusive nor exhaustive.
- Clarifies the nature of the problem to be solved
- Can be used to suggest or generate hypotheses
- Includes the use of pilot studies
- Used widely in market research
- Provides general frequency data about populations or samples
- Does not manipulate variables (e.g. as in an experiment)
- Describes only the “who, what, when, where and how”
- Cannot establish a causal relationship between variables
- Associated with descriptive statistics
- Breaks down factors or variables involved in a concept, problem or issue
- Often uses (or generates) models as analytical tools
- Often uses micro/macro distinctions in analysis
- Focuses on the analysis of bias, inconsistencies, gaps or contradictions in accounts, theories, studies or models
- Often takes a specific theoretical perspective, (e.g. feminism; labour process theory)
- Mainly quantitative
- Identifies measurable variables
- Often manipulates variables to produce measurable effects
- Uses specific, predictive or null hypotheses
- Dependent on accurate sampling
- Uses statistical testing to establish causal relationships, variance between samples or predictive trends
- Associated with organisation development initiatives and interventions
- Practitioner based, works with practitioners to help them solve their problems
- Involves data collection, evaluation and reflection
- Often used to review interventions and plan new ones
- Focuses on recognised needs, solving practical problems or answering specific questions
- Often has specific commercial objectives (e.g. product development )
Approaches to research
For many, perhaps most, researchers, the choice of approach is straightforward. Research into reaction mechanisms for an organic chemical reaction will take a quantitative approach, whereas qualitative research will have a better fit in the social work field that focuses on families and individuals. While some research benefits from one of the two approaches, other research yields more understanding from a combined approach.
In fact, qualitative and quantitative approaches to research have some important shared aspects. Each type of research generally follows the steps of scientific method, specifically:
In general, each approach begins with qualitative reasoning or a hypothesis based on a value judgement. These judgements can be applied, or transferred to quantitative terms with both inductive and deductive reasoning abilities. Both can be very detailed, although qualitative research has more flexibility with its amount of detail.
Selecting an appropriate design for a study involves following a logical thought process; it is important to explore all possible consequences of using a particular design in a study. As well as carrying out a scoping study, a researchers should familiarise themselves with both qualitative and quantitative approaches to research in order to make the best decision. Some researchers may quickly select a qualitative approach out of fear of statistics but it may be a better idea to challenge oneself. The researcher should also be prepared to defend the paradigm and chosen research method; this is even more important if your proposal or grant is for money, or other resources.
Ultimately, clear goals and objectives and a fit-for-purpose research design is more helpful and important than old-fashioned arguments about which approach to research is “best”. Indeed, there is probably no such thing as a single “correct” design – hypotheses can be studied by different methods using different research designs. A research design is probably best thought of as a series of signposts to keep the research headed in the right direction and should not be regarded as a highly specific plan to be followed without deviation.
Research models
There is no common agreement on the classification of research models but, for the purpose of illustration, five categories of research models and their variants are outlined below.
A physical model is a physical object shaped to look like the represented phenomenon, usually built to scale e.g. atoms, molecules, skeletons, organs, animals, insects, sculptures, small-scale vehicles or buildings, life-size prototype products. They can also include 3-dimensional alternatives for two-dimensional representations e.g. a physical model of a picture or photograph.
In this case, the term model is used loosely to refer to any theory phrased in formal, speculative or symbolic styles. They generally consist of a set of assumptions about some concept or system; are often formulated, developed and named on the basis of an analogy between the object, or system that it describes and some other object or different system; and they are considered an approximation that is useful for certain purposes. Theoretical models are often used in biology, chemistry, physics and psychology.
A mathematical model refers to the use of mathematical equations to depict relationships between variables, or the behaviour of persons, groups, communities, cultural groups, nations, etc.
It is an abstract model that uses mathematical language to describe the behaviour of a system. They are used particularly in the natural sciences and engineering disciplines (such as physics, biology, and electrical engineering) but also in the social sciences (such as economics, sociology and political science). Types of mathematical models include trend (time series), stochastic, causal and path models. Examples include models of population and economic growth, weather forecasting and the characterisation of large social networks.
Mechanical (or computer) models tend to use concepts from the natural sciences, particularly physics, to provide analogues for social behaviour. They are often an extension of mathematical models. Many computer-simulation models have shown how a research problem can be investigated through sequences of experiments e.g. game models; microanalytic simulation models (used to examine the effects of various kinds of policy on e.g. the demographic structure of a population); models for predicting storm frequency, or tracking a hurricane.
These models are used to untangle meanings that individuals give to symbols that they use or encounter. They are generally simulation models, i.e. they are based on artificial (contrived) situations, or structured concepts that correspond to real situations. They are characterised by symbols, change, interaction and empiricism and are often used to examine human interaction in social settings.
The advantages and disadvantages of modelling
Take a look at the advantages and disadvantages below. It might help you think about what type of model you may use.
- The determination of factors or variables that most influence the behaviour of phenomena
- The ability to predict, or forecast the long term behaviour of phenomena
- The ability to predict the behaviour of the phenomenon when changes are made to the factors influencing it
- They allow researchers a view on difficult to study processes (e.g. old, complex or single-occurrence processes)
- They allow the study of mathematically intractable problems (e.g. complex non-linear systems such as language)
- They can be explicit, detailed, consistent, and clear (but that can also be a weakness)
- They allow the exploration of different parameter settings (i.e. evolutionary, environmental, individual and social factors can be easily varied)
- Models validated for a category of systems can be used in many different scenarios e.g. they can be reused in the design, analysis, simulation, diagnosis and prediction of a technical system
- Models enable researchers to generate unrealistic scenarios as well as realistic ones
- Difficulties in validating models
- Difficulties in assessing the accuracy of models
- Models can be very complex and difficult to explain
- Models do not “provide proof”
The next section describes the processes and design of research.
No internet connection.
All search filters on the page have been cleared., your search has been saved..
- Sign in to my profile My Profile
Welcome to Sage Research Methods
Navigating away from this page will delete your results
Please save your results to "My Self-Assessments" in your profile before navigating away from this page.
Sign in to my profile
Please sign into your institution before accessing your profile
Sign up for a free trial and experience all Sage Learning Resources have to offer.
You must have a valid academic email address to sign up.
Get off-campus access
- View or download all content my institution has access to.
Sign up for a free trial and experience all Sage Learning Resources has to offer.
- view my profile
- view my lists
Researching and Developing Models, Theories and Approaches for Design and Development
- First Online: 15 November 2023
Cite this chapter
- David C. Wynn 3 &
- P. John Clarkson 4
459 Accesses
This chapter discusses the research-driven development of models, theories and approaches for design and development. It begins by clarifying the types of models, theories and approaches considered. Desirable characteristics for each specific type are then outlined, and research methods for developing and evaluating them are discussed. A framework is introduced to organise these methodological considerations.
This is a preview of subscription content, log in via an institution to check access.
Access this chapter
Subscribe and save.
- Get 10 units per month
- Download Article/Chapter or eBook
- 1 Unit = 1 Article or 1 Chapter
- Cancel anytime
- Available as PDF
- Read on any device
- Instant download
- Own it forever
- Available as EPUB and PDF
- Durable hardcover edition
- Dispatched in 3 to 5 business days
- Free shipping worldwide - see info
Tax calculation will be finalised at checkout
Purchases are for personal use only
Institutional subscriptions
Ahmad, N., Wynn, D. C., & Clarkson, P. J. (2013). Change impact on a product and its redesign process: A tool for knowledge capture and reuse. Research in Engineering Design 24 (3), 219–244. https://doi.org/10.1007/s00163-012-0139-8
Antonsson, E. K. (1987). Development and testing of hypotheses in engineering design research. Journal of Mechanisms, Transmissions, and Automation in Design, 109 (2), 153–154. https://doi.org/10.1115/1.3267429
Araujo, C. S., Benedetto-Neto, H., Campello, A. C., Segre, F. M., & Wright, I. C. (1996). The utilization of product development methods: A survey of UK industry. Journal of Engineering Design, 7 (3), 265–277. https://doi.org/10.1080/09544829608907940
Bacharach, S. B. (1989). Organizational theories: Some criteria for evaluation. Academy of Management Review, 14 (4), 496–515. https://doi.org/10.5465/amr.1989.4308374
Barth, A., Caillaud, E., & Rose, B. (2011). How to validate research in engineering design? In S. J. Culley, B. J. Hicks, T. C. McAloone, T. J. Howard, & Y. Reich (Eds.), DS 68-2: Proceedings of the 18th International Conference on Engineering Design (ICED 11), Impacting Society through Engineering Design, Vol. 2: Design Theory and Research Methodology, Lyngby/Copenhagen, Denmark. The Design Society.
Google Scholar
Blessing, L. T. M., & Chakrabarti, A. (2009). DRM, a design research methodology . London: Springer. https://doi.org/10.1007/978-1-84882-587-1
Bracewell, R. H., Shea, K., Langdon, P. M., Blessing, L. T. M., & Clarkson, P. J. (2001). A methodology for computational design tool research. In Proceedings of ICED01, Glasgow, Scotland , pp. 181–188.
Bracewell, R., Wallace, K., Moss, M., & Knott, D. (2009). Capturing design rationale. Computer-Aided Design, 41 (3), 173–186. https://doi.org/10.1016/j.cad.2008.10.005
Cash, P., Isaksson, O., Maier, A., & Summers, J. (2022). Sampling in design research: Eight key considerations. Design Studies, 78 , 101077. https://doi.org/10.1016/j.destud.2021.101077
Cash, P. J. (2018). Developing theory-driven design research. Design Studies, 56 , 84–119. https://doi.org/10.1016/j.destud.2018.03.002
Cooper, H. M. (1988). Organizing knowledge syntheses: A taxonomy of literature reviews. Knowledge in Society, 1 , 104. https://doi.org/10.1007/BF03177550
Daalhuizen, J., & Cash, P. (2021). Method content theory: Towards a new understanding of methods in design. Design Studies, 75 , 101018. https://doi.org/10.1016/j.destud.2021.101018
Dixon, J. R. (1987). On research methodology towards a scientific theory of engineering design. Artificial Intelligence for Engineering Design, Analysis and Manufacturing, 1 (3), 145–157. https://doi.org/10.1017/S0890060400000251
Eckert, C. M., Stacey, M. K., & Clarkson, P. J. (2003). The spiral of applied research: A method ological view on integrated design research. In A. Folkeson, K. Gralen, M. Norell, & U. Sellgren (Eds.), DS 31: Proceedings of ICED 03, the 14th International Conference on Engineering Design, Stockholm (pp. 245–246). The Design Society.
Eisenhardt, K. M. (1989). Building theories from case study research. Academy of Management Review, 14 (4), 532–550. https://doi.org/10.2307/258557
Ericsson, K. A. (2017). Protocol analysis. In W. Bechtel & G. Graham (Eds.), A companion to cognitive science (pp. 425–432). Hoboken: Wiley. https://doi.org/10.1002/9781405164535.ch33
Fowler, F. J. J., & Floyd, J. (2013). Survey research methods . Thousand Oaks, CA: Sage Publications.
Frey, D. D., & Dym, C. L. (2006). Validation of design methods: Lessons from medicine. Research in Engineering Design, 17 (1), 45–57. https://doi.org/10.1007/s00163-006-0016-4
Gabaix, X., & Laibson, D. (2008). The seven properties of good models. In A. Caplin & A. Schotter (Eds.), The foundations of positive and normative economics: A handbook (pp. 292–319). New York: Oxford University Press. https://doi.org/10.1093/acprof:oso/9780195328318.003.0012
Chapter Google Scholar
Gericke, K., Eckert, C., & Stacey, M. (2017). What do we need to say about a design method? In A. Maier, S. Škec, H. Kim, M. Kokkolaras, J. Oehmen, G. Fadel, F. Salustri, & M. V. der Loos (Eds.), DS 87-7 Proceedings of the 21st International Conference on Engineering Design (ICED 17) Vol 7: Design Theory and Research Methodology, Vancouver, Canada (pp. 101–110). The Design Society.
Glaser, B. G., & Strauss, A. L. (1999). The discovery of grounded theory: Strategies for qualitative research . New York: Routledge. https://doi.org/10.4324/9780203793206
Book Google Scholar
Guba, E. G. (1981). Criteria for assessing the trustworthiness of naturalistic inquiries. ECTJ, 29 (2), 75–91. https://doi.org/10.1007/BF02766777
Article Google Scholar
Gubrium, J. F., & Holstein, J. A. (2001). Handbook of interview research: Context and method . Thousand Oaks, CA: Sage Publications. https://doi.org/10.4135/9781412973588
Hay, L., Duffy, A. H. B., McTeague, C., Pidgeon, L. M., Vuletic, T., & Grealy, M. (2017). Towards a shared ontology: A generic classification of cognitive processes in conceptual design. Design Science, 3 , e7. https://doi.org/10.1017/dsj.2017.6
Isaksson, O., Eckert, C., Panarotto, M., & Malmqvist, J. (2020). You need to focus to validate. Proceedings of the Design Society: DESIGN Conference, 1 , 31–40. https://doi.org/10.1017/dsd.2020.116
Jabareen, Y. (2009). Building a conceptual framework: Philosophy, definitions, and procedure. International Journal of Qualitative Methods, 8 (4), 49–62. https://doi.org/10.1177/160940690900800406
Kerley, W. P., Wynn, D. C., Moss, M. A., Coventry, G., & Clarkson, P. J. (2009). Towards empirically-derived guidelines for process modelling interventions in engineering design. In M. Norell Bergendahl, M. Grimheden, L. Leifer, P. Skogstad, & U. Lindemann (Eds.), DS 58-1: Proceedings of ICED 09, the 17th International Conference on Engineering Design, Vol. 1, Design Processes, Palo Alto, CA, USA (pp. 217–228). The Design Society.
Lavrsen, J. C., Daalhuizen, J., Dømler, S., & Fisker, K. (2022). Towards a lifecycle of design methods. In D. Lockton, S. Lenzi, P. Hekkert, A. Oak, J. Sádaba, & P. Lloyd (Eds.), DRS2022: Bilbao, 25 June - 3 July. Spain: Bilbao. https://doi.org/10.21606/drs.2022.542
Law, A. M. (2008). How to build valid and credible simulation models. In S. J. Mason, R. R. Hill, L. Mönch, O. Rose, T. Jefferson, & J. W. Fowler (Eds.), Proceedings of the 2008 Winter Simulation Conference (pp. 39–47). IEEE. https://doi.org/10.1109/WSC.2008.4736054
Le Dain, M.-A., Blanco, E., & Summers, J. D. (2013). Assessing design research quality: In vestigating verification and validation criteria. In U. Lindemann, S. Venkataraman, Y. S. Kim, S. W. Lee, Y. Reich, & A. Chakrabarti (Eds.), DS 75-2: Proceedings of the 19th International Conference on Engineering Design (ICED13), Design for Harmonies, Vol. 2: Design Theory and Research Methodology, Seoul, Korea, 19-22.08. 2013 (pp. 183–192). The Design Society.
Levy, S., Subrahmanian, E., Konda, S., Coyne, R., Westerberg, A., & Reich, Y. (1993). An overview of the n-dim environment (Technical Report EDRC-05-65-93). Engineering Design Research Center, Carnegie-Mellon University, Pittsburgh.
Li, Y., Horváth, I., & Rusák, Z. (2022). An underpinning theory and approach to applicability testing of constructive computational mechanisms. Research in Engineering Design, 33 (2), 213–230. https://doi.org/10.1007/s00163-022-00385-0
Little, J. D. (1970). Models and managers: The concept of a decision calculus. Management Science, 16 (8), B-466-B-485. https://doi.org/10.1287/mnsc.16.8.B466
Moody, D. L. (2005). Theoretical and practical issues in evaluating the quality of conceptual models: Current state and future directions. Data & Knowledge Engineering, 55 (3), 243–276. https://doi.org/10.1016/j.datak.2004.12.005
Olewnik, A. T., & Lewis, K. (2005). On validating engineering design decision support tools. Concurrent Engineering Research and Applications, 13 (2), 111–122. https://doi.org/10.1177/1063293X05053796
Pedersen, K., Emblemsvåg, J., Bailey, R., Allen, J. K., & Mistree, F. (2000). Validating design methods and research: The validation square. In Proceedings of the ASME 2000 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. Volume 4: 12th International Conference on Design Theory and Methodology. Baltimore, Maryland, USA. September 10–13, 2000 , pp. 379–390. https://doi.org/10.1115/DETC2000/DTM-14579
Pidd, M. (1999). Just modeling through: A rough guide to modeling. Interfaces, 29 (2), 118–132. https://doi.org/10.1287/inte.29.2.118
Reich, Y. (1994). Layered models of research methodologies. Artificial Intelligence for Engineering Design, Analysis and Manufacturing, 8 (4), 263–274. https://doi.org/10.1017/S0890060400000949
Reich, Y. (2017). The principle of reflexive practice. Design Science, 3 , e4. https://doi.org/10.1017/dsj.2017.3
Reich, Y., Konda, S., Subrahmanian, E., Cunningham, D., Dutoit, A., Patrick, R., Thomas, M., & Westerberg, A. W. (1999). Building agility for developing agile design information systems. Research in Engineering Design, 11 (2), 67–83. https://doi.org/10.1007/PL00003884
Robinson, S. (2008). Conceptual modelling for simulation part I: Definition and requirements. Journal of the Operational Research Society, 59 , 278–290. https://doi.org/10.1057/palgrave.jors.2602368
Sargent, R. G. (2013). Verification and validation of simulation models. Journal of Simulation, 7 (1), 12–24. https://doi.org/10.1057/jos.2012.20
Seepersad, C. C., Pedersen, K., Emblemsvåg, J., Bailey, R., Allen, J. K., & Mistree, F. (2006). The Validation Square: How Does One Verify and Validate a Design Method? In K. E. Lewis, W. Chen, & L. C. Schmidt (Eds.), Decision Making in Engineering Design. New York: ASME Press. https://doi.org/10.1115/1.802469.ch25
Snyder, H. (2019). Literature review as a research methodology: An overview and guidelines. Journal of Business Research, 104 , 333–339. https://doi.org/10.1016/j.jbusres.2019.07.039
Teegavarapu, S., Summers, J. D., & Mocko, G. M. (2008). Case Study Method for Design Research: A Justification. In Proceedings of the ASME 2008 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. Volume 4: 20th International Conference on Design Theory and Methodology; Second International Conference on Micro- and Nanosystems. Brooklyn, New York, USA. August 3–6, 2008 (pp. 495–503). https://doi.org/10.1115/DETC2008-49980
Torraco, R. J. (2005). Writing integrative literature reviews: Guidelines and examples. Human Resource Development Review, 4 (3), 356–367. https://doi.org/10.1177/1534484305278283
Torraco, R. J. (2016). Writing integrative literature reviews: Using the past and present to explore the future. Human Resource Development Review, 15 (4), 404–428. https://doi.org/10.1177/1534484316671606
Van der Waldt, G. (2020). Constructing conceptual frameworks in social science research. The Journal for Transdisciplinary Research in Southern Africa, 16 (1), 1–9. https://doi.org/10.4102/td.v16i1.758
Wacker, J. G. (2008). A conceptual understanding of requirements for theory-building research: Guidelines for scientific theory building. Journal of Supply Chain Management, 44 (3), 5–15. https://doi.org/10.1111/j.1745-493X.2008.00062.x
Wallace, K. (2011). Transferring design methods into practice. In H. Birkhofer (Ed.), The future of design methodology (pp. 239–248). London: Springer. https://doi.org/10.1007/978-0-85729-615-3_21
Wyatt, D. F., Wynn, D. C., Jarrett, J. P., & Clarkson, P. J. (2012). Supporting product architecture design using computational design synthesis with network structure constraints. Research in Engineering Design, 23 (1), 17–52. https://doi.org/10.1007/s00163-011-0112-y
Wynn, D. C., Caldwell, N. H. M., & Clarkson, P. J. (2014). Predicting change propagation in complex design workflows. Journal of Mechanical Design, 136 (8), 081009. https://doi.org/10.1115/1.4027495
Wynn, D. C., Wyatt, D. F., Nair, S. M. T., & Clarkson, P. J. (2010). An introduction to the Cambridge Advanced Modeller. In P. Heisig, P. J. Clarkson, & S. Vajna (Eds.), Proceedings of the 1st International Conference on Modelling and Management of Engineering Processes (MMEP 2010). Cambridge, UK, 19–20 July 2010.
Wynn, D. C., & Clarkson, P. J. (2021). Improving the engineering design process by simulating iteration impact with ASM2.0. Research in Engineering Design, 32 (2), 127–156. https://doi.org/10.1007/s00163-020-00354-5
Yin, R. K. (2017). Case study research and applications: Design and methods (6th ed.). Los Angeles: Sage.
Download references
Author information
Authors and affiliations.
Department of Mechanical and Mechatronics Engineering, University of Auckland, Auckland, New Zealand
David C. Wynn
Department of Engineering, University of Cambridge, Cambridge, UK
P. John Clarkson
You can also search for this author in PubMed Google Scholar
Corresponding author
Correspondence to David C. Wynn .
Rights and permissions
Reprints and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Wynn, D.C., Clarkson, P.J. (2024). Researching and Developing Models, Theories and Approaches for Design and Development. In: The Design and Development Process. Springer, Cham. https://doi.org/10.1007/978-3-031-38168-3_5
Download citation
DOI : https://doi.org/10.1007/978-3-031-38168-3_5
Published : 15 November 2023
Publisher Name : Springer, Cham
Print ISBN : 978-3-031-38167-6
Online ISBN : 978-3-031-38168-3
eBook Packages : Engineering Engineering (R0)
Share this chapter
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
- Publish with us
Policies and ethics
- Find a journal
- Track your research
Step 2: Pick a theory, model, and/or framework
One of the cornerstones of implementation science is the use of theory. These theories, models, and frameworks (TMFs) provide a structured approach to understanding, guiding, and evaluating the complex process of translating research into practical applications, ultimately improving healthcare and other professional practices.
Theories, models, and frameworks serve several critical functions in implementation science. They help researchers and practitioners comprehend the multifaceted nature of implementation processes, including the factors that influence the adoption, implementation, and sustainability of interventions. TMFs offer structured pathways and strategies for planning and executing implementation efforts, ensuring that interventions are systematically and effectively integrated into practice. Additionally, they provide criteria and methods for assessing the success of implementation efforts, identifying barriers and facilitators, and informing continuous improvement. To learn more about the use of theory in implementation science, read Harnessing the power of theorising in implementation science (Kislov et al, 2019) and Theorizing is for everybody: Advancing the process of theorizing in implementation science (Meza et al, 2023)
There are dozens of TMFs used in implementation science, developed across a wide range of disciplines such as psychology, sociology, organizational theory, and public health. Some well-known examples include the Consolidated Framework for Implementation Research (CFIR), which identifies constructs across five domains that can influence implementation outcomes; the Exploration, Preparation, Implementation, Sustainment (EPIS) Framework, which emphasizes the importance of involving stakeholders at all levels and stages of the implementation process; and the Promoting Action on Research Implementation in Health Services (PARIHS) Framework, which focuses on the interplay between evidence, context, and facilitation in successful implementation.
The vast number of available TMFs can make it challenging to determine which is the most appropriate to address or frame a research question. Two notable reviews provide schemas to organize and narrow the range of choices. Nilsen (2015) categorizes TMFs into five types: process models, determinants frameworks, classic theories, implementation theories, and evaluation frameworks. Tabak et al. (2013) organizes 61 dissemination and implementation models based on construct flexibility, focus on dissemination and/or implementation activities, and socio-ecological framework level.
✪ Making sense of implementation theories, models, and frameworks
(Nilsen, 2015)
Theoretical approaches enhance our understanding of why implementation succeeds or fails, so by considering both process models and determinant frameworks, researchers and practitioners can improve the implementation of evidence-based practices across various contexts.
Nilsen's schema sorts implementation science theories, models, and frameworks into five categories:
- Process models: These describe or guide the translation of research evidence into practice, outlining the steps involved in implementing evidence-based practices.
- Determinants frameworks: These focus on understanding and explaining the factors that influence implementation outcomes, highlighting barriers and enablers (but may lack specific practical guidance).
- Classic theories: These are established theories from various disciplines (e.g., psychology, sociology) that inform implementation.
- Implementation theories: These have been specifically designed to address implementation processes and outcomes.
- Evaluation frameworks: These assess the effectiveness of implementation efforts, helping evaluate whether the intended changes have been successfully implemented.
While there is some overlap between these theories, models, and frameworks, understanding their differences is essential for selecting relevant approaches in research and practice.
Adapted from: Nilsen P. Making sense of implementation theories, models and frameworks. Implement Sci . 2015;10(1):1-13.
Bridging research and practice: models for dissemination and implementation research
(Tabak, Khoong, Chambers, & Brownson, 2013)
Rachel Tabak and colleagues organized 61 dissemination and implementation theories, models, and frameworks (TMFs) based on three variables to provide researchers with a better understanding of how these tools can be best used in dissemination and implementation (D&I) research.
These variables are:
- Construct flexibility: Some are broad and flexible, allowing for adaptation to different contexts, while others are more specific and operational
- Focus on Dissemination or Implementation Activities: These range from being dissemination-focused (emphasizing the spread of information) to implementation-focused (emphasizing the adoption and integration of interventions)
- Socioecologic Framework level: Addressing system, community, organization, or individual levels, with fewer models addressing policy activities
The authors argue that classification of a TMF based on these three variables will assist in selecting a TMF best suited to inform dissemination and implementation science study design and execution and answer the question being asked.
For more information, watch: 💻 Applying Models and Frameworks to D&I Research: An Overview & Analysis , with presenters Dr. Rachel Tabak and Dr. Ted Skolarus.
The IS Research Pathway
- 1. Frame Your Question
- 2. Pick Theories, Models, & Frameworks
- 3. Select Implementation Strategies
- 4. Select Research Method
- 5. Choose Study Design
- 6. Choose Measures
- 7. Get Funding
- 8. Report Results
🎥 Videos from our friends
Dr. charles jonassaint, university of pittsburgh dissemination and implementation science collaborative, dr. rachel shelton, scd, mph & dr. nathalie moise, md, ms, faha, columbia university, dr. meghan lane-fall, md, mshp, fccm, penn implementation science center (pisce).
✪ Using implementation science theories and frameworks in global health ( BMJ Global Health , 2019)
✪ A scoping review of implementation science theories, models, and frameworks — an appraisal of purpose, characteristics, usability, applicability, and testability ( Implementation Science , 2023)
✪ Ten recommendations for using implementation frameworks in research and practice ( Implementation Science Communications , 2020)
Understanding implementation science from the standpoint of health organisation and management: an interdisciplinary exploration of selected theories, models and frameworks ( Journal of Health Organization and Management , 2021)
💻 The D&I Models Webtool
💻 Toolkit Part 1: Implementation Science Methodologies and Frameworks ( Fogarty International Center )
🎥 Implementation Science Theories, Frameworks, and Models ( UAB CFAR Implementation Science Hub )
🎥 Theories and Frameworks in Implementation Science ( Interdisciplinary Research Leaders )
A Selection of TMFs
While both ways of viewing this array of tools are useful, below we borrow from Nilsen’s schema to organize overviews of a selection of implementation science theories, models, and frameworks. In each overview, you will find links to additional resources.
Open Access articles will be marked with ✪ Please note some journals will require subscriptions to access a linked article.
Process models, used to describe or guide the process of translating research into practice, dynamic adaptation process.
The need to adapt to local context is a consistent theme in the adoption of evidence-based practices, and Aarons and colleagues created the Dynamic Adaption Process to address this need. Finding that adaptation was often ad hoc rather than intentional and planned, the Dynamic Adaption Process helps identify core components and adaptable characteristics of an evidence-based practice and supports implementation with training. The result of the Dynamic Adaption Process is a data-informed, collaborative, and stakeholder-driven approach to maintaining intervention fidelity during evidence-based practice implementation, addressing real-world implications for public sector service systems and is relevant at national, state, and local levels. The framework development article, Dynamic adaptation process to implement an evidence-based child maltreatment intervention , was published in 2012 in the open access journal, Implementation Science .
Examples of Use
- ✪ “Scaling-out” evidence-based interventions to new populations or new health care delivery systems ( Implementation Science , 2017)
- ✪ Implementing measurement-based care in community mental health: a description of tailored and standardized methods ( Implementation Science , 2018)
- ✪ “I Had to Somehow Still Be Flexible”: Exploring Adaptations During Implementation of Brief Cognitive Behavioral Therapy in Primary Care ( Implementation Science , 2018)
- An Implementation Science Approach to Antibiotic Stewardship in Emergency Departments and Urgent Care Centers ( Academic Emergency Medicine , 2020)
- ✪ Initial adaptation of the OnTrack coordinated specialty care model in Chile: An application of the Dynamic Adaptation Process ( Frontiers in Health Services , 2022)
Exploration, Adoption/Preparation, Implementation, Sustainment Model (EPIS)
Recognizing that implementation science frameworks were largely developed using research from business and medical contexts, Aarons, Hurlburt, and Horwitz created the four-phase implementation model EPIS (Exploration, Adoption/Preparation, Implementation, Sustainment) in 2010 to address implementation in public service sector contexts. The EPIS framework offers a systematic approach to understanding and implementing evidence-based practices, considering context, and ensuring sustainability throughout the process. The framework development article, Advancing a Conceptual Model of Evidence-Based Practice Implementation in Public Service Sectors , is available open access (✪) from Administration and Policy in Mental Health and Mental Health Services Research . You can also learn more by visiting EPISFramework.com .
In 2018 the authors refined the EPIS model into the cyclical EPIS Wheel, allowing for closer alignment with rapid-cycle testing. A model for rigorously applying the Exploration, Preparation, Implementation, Sustainment (EPIS) framework in the design and measurement of a large-scale collaborative multi-site study is available Open Access (✪) from Health & Justice .
- ✪ Systematic review of the Exploration, Preparation, Implementation, Sustainment (EPIS) framework ( Implementation Science , 2019)
- A Review of Studies on the System-Wide Implementation of Evidence-Based Psychotherapies for Posttraumatic Stress Disorder in the Veterans Health Administration ( Administration and Policy in Mental Health and Mental Health Services Research , 2016)
- Advancing Implementation Research and Practice in Behavioral Health Systems ( Administration and Policy in Mental Health and Mental Health Services Research , 2016)
- ✪ A model for rigorously applying the Exploration, Preparation, Implementation, Sustainment (EPIS) framework in the design and measurement of a large scale collaborative multi-site study ( Health and Justice , 2018)
- Characterizing Shared and Unique Implementation Influences in Two Community Services Systems for Autism: Applying the EPIS Framework to Two Large-Scale Autism Intervention Community Effectiveness Trials ( Administration and Policy in Mental Health and Mental Health Services Research , 2019)
- ✪ ✪ Systematic review of the Exploration, Preparation, Implementation, Sustainment (EPIS) framework ( Implementation Science , 2019)
- ✪ The core functions and forms paradigm throughout EPIS: designing and implementing an evidence-based practice with function fidelity ( Frontiers in Health Services , 2024)
- 💻 WEBINAR: Use of theory in implementation research: The EPIS framework: A phased and multilevel approach to implementation
Quality Implementation Framework
In 2012 Meyers, Durlak, and Wandersman synthesized information from 25 implementation frameworks with a focus on identifying specific actions that improve the quality of implementation efforts. The result of this synthesis was the Quality Implementation Framework (QIF) , published in the American Journal of Community Psychology . The QIF provides a roadmap for achieving successful implementation by breaking down the process into actionable steps across four phases of implementation:
- Exploration: In this phase, stakeholders explore and consider the need for implementing a specific intervention
- Installation: This phase involves planning and preparing for implementation
- Initial Implementation: The intervention is put into action during this phase
- Full Implementation: The focus shifts to sustaining the intervention over time
- ✪ Practical Implementation Science: Developing and Piloting the Quality Implementation Tool ( American Journal of Community Psychology , 2012)
- Survivorship Care Planning in a Comprehensive Cancer Center Using an Implementation Framework ( The Journal of Community and Supportive Oncology , 2016)
- ✪ The Application of an Implementation Science Framework to Comprehensive School Physical Activity Programs: Be a Champion! ( Frontiers in Public Health , 2017)
- ✪ Developing and Evaluating a Lay Health Worker Delivered Implementation Intervention to Decrease Engagement Disparities in Behavioural Parent Training: A Mixed Methods Study Protocol ( BMJ Open , 2019)
- Implementation Process and Quality of a Primary Health Care System Improvement Initiative in a Decentralized Context: A Retrospective Appraisal Using the Quality Implementation Framework ( The International Journal of Health Planning and Management , 2019)
Determinant Frameworks
Used to understand and/or explain what influences implementation outcomes, active implementation framework.
In 2005, the National Implementation Research Network (NIRN) published an Open Access (✪) monograph synthesizing transdisciplinary research on implementation evaluation. The resulting Active Implementation Frameworks (AIFs) include the following five elements: Usable Intervention Criteria, Stages of implementation, Implementation Drivers, Improvement Cycles, and Implementation Teams. A robust support and training website is maintained by NIRN, complete with activities and assessments to guide active implementation.
Learn More:
- Statewide Implementation of Evidence-Based Programs ( Exceptional Children , 2013)
- Active Implementation Frameworks for Successful Service Delivery: Catawba County Child Wellbeing Project ( Research on Social Work Practice , 2014)
- The Active Implementation Frameworks: A roadmap for advancing implementation of Comprehensive Medication Management in primary care ( Research in Social and Administrative Pharmacy , 2017)
Consolidated Framework for Implementation Research (CFIR)
In 2009, Veterans Affairs researchers developed a menu of constructs found to be associated with effective implementation across 13 scientific disciplines. Their goal was to review the wide range of terminology and varying definitions used in implementation research, then construct an organizing framework that considered them all. The resulting Consolidated Framework for Implementation Research (CFIR) has been widely cited and has been found useful across a range of disciplines in diverse settings. Designed to guide the systematic assessment of multilevel implementation contexts, the CFIR helps identify factors that might influence the implementation and effectiveness of interventions. The CFIR provides a menu of constructs associated with effective implementation, reflecting the state-of-the-science at the time of its development in 2009. By offering a framework of constructs, the CFIR promotes consistent use, systematic analysis, and organization of findings from implementation studies. In 2022, the CFIR was updated based on feedback from CFIR users, addressing critiques by updating construct names and definitions, adding missing constructs, and dividing existing constructs for needed nuance. A CFIR Outcomes Addendum was also published in 2022, to offer clear conceptual distinctions between types of outcomes for use with the CFIR, helping bring clarity as researchers think about which outcomes are most appropriate for their research question.
For additional resources, please visit the CFIR Technical Assistance Website . The website has tools and templates for studying implementation of innovations using the CFIR framework, and these tools can help you learn more about issues pertaining to inner and outer contexts. You can read the original framework development article in the Open Access (✪) journal Implementation Science .
- ✪ Evaluating and Optimizing the Consolidated Framework for Implementation Research (CFIR) for use in Low- and Middle-Income Countries: A Systematic Review ( Implementation Science , 2020)
- ✪ A systematic review of the use of the Consolidated Framework for Implementation Research ( Implementation Science , 2017)
- Using the Consolidated Framework for Implementation Research (CFIR) to produce actionable findings: A rapid-cycle evaluation approach to improving implementation ( Implementation Science , 2017)
- ✪ The Consolidated Framework for Implementation Research: Advancing implementation science through real-world applications, adaptations, and measurement ( Implementation Science , 2015)
- 💻 WEBINAR: Use of theory in implementation research: Pragmatic application and scientific advancement of the Consolidated Framework for Implementation Research (CFIR) (Dr. Laura Damschroder, National Cancer Institute of NIH Fireside Chat Series )
- 💻 Updated CFIR, Explained (University of Pittsburgh DISC)
Dynamic Sustainability Framework
The Dynamic Sustainability Framework arose from the need to better understand how the sustainability of health interventions can be improved. In 2013, Chambers, Glasgow, and Stange published ✪ The dynamic sustainability framework: addressing the paradox of sustainment amid ongoing change in the Open Access (✪) journal Implementation Science . While traditional models of sustainability often assume diminishing benefits over time, this framework challenges those assumptions. It emphasizes continuous learning, problem-solving, and adaptation of interventions within multi-level contexts. Rather than viewing sustainability as an endgame, the framework encourages ongoing improvement and integration of interventions into local organizational and cultural contexts. By focusing on fit between interventions and their changing context, the Dynamic Sustainability Framework aims to advance the implementation, transportability, and impact of health services research.
- ✪ Results-based aid with lasting effects: Sustainability in the Salud Mesoamérica Initiative ( Globalization and Health , 2018)
- ✪ Study Protocol: A Clinical Trial for Improving Mental Health Screening for Aboriginal and Torres Strait Islander Pregnant Women and Mothers of Young Children Using the Kimberley Mum’s Mood Scale ( BMC Public Health , 2019)
- ✪ Sustainability of Public Health Interventions: Where Are the Gaps? ( Health Research Policy and Systems , 2019)
Practical, Robust Implementation and Sustainability Model (PRISM)
In 2008, Feldstein and Glasgow developed the Practical, Robust Implementation and Sustainability Model (PRISM) to address the lack of consideration of non-research settings in efficacy and effectiveness trials. This model evaluates how the intervention interacts with recipients to influence program adoption, implementation, maintenance, reach, and effectiveness. The framework development article was published by The Joint Commission Journal on Quality and Patient Safety . In 2022, Rabin and colleagues published a follow up article, ‘ A citation analysis and scoping systematic review of the operationalization of the Practical, Robust Implementation and Sustainability Model (PRISM) ‘ which aimed to assess the use of the PRISM framework and to make recommendations for future research.
- Using the Practical, Robust Implementation and Sustainability Model (PRISM) to Qualitatively Assess Multilevel Contextual Factors to Help Plan, Implement, Evaluate, and Disseminate Health Services Programs ( Translational Behavioral Medicine , 2019)
- Stakeholder Perspectives on Implementing a Universal Lynch Syndrome Screening Program: A Qualitative Study of Early Barriers and Facilitators ( Genetics Medicine , 2016)
- Evaluating the Implementation of Project Re-Engineered Discharge (RED) in Five Veterans Health Administration (VHA) Hospitals ( The Joint Commission Journal on Quality and Patient Safety , 2018)
- ✪ Applying an equity lens to assess context and implementation in public health and health services research and practice using the PRISM framework ( Frontiers in Health Services , 2023)
- ✪ Integrating the Practical Robust Implementation and Sustainability Model With Best Practices in Clinical Decision Support Design: Implementation Science Approach ( Journal of Medical Internet Research , 2020)
Promoting Action on Research Implementation in Health Services (PARIHS) Framework
Using collective experience in research, practice development, and quality improvement efforts, Kitson, Harvey and McCormack proposed in 1998 that success in implementation is a result of the interactions between evidence, context, and facilitation. Their resulting Promoting Action on Research Implementation in Health Services (PARIHS) framework posits that successful implementation requires clear understanding of the evidence in use, the context involved, and the type of facilitation utilized to achieve change.
The original framework development article, Enabling the implementation of evidence based practice: a conceptual framework is available Open Access (✪) from BMJ Quality & Safety .
- Ingredients for change: revisiting a conceptual framework ( BMJ Quality & Safety , 2002)
- Evaluating the successful implementation of evidence into practice using the PARIHS framework: theoretical and practical challenges ( Implementation Science , 2008)
- ✪ A critical synthesis of literature on the promoting action on research implementation in health services (PARIHS) framework ( Implementation Science , 2010)
- ✪ A Guide for applying a revised version of the PARIHS framework for implementation ( Implementation Science , 2011)
- 💻 WEBINAR: Use of theory in implementation research; Pragmatic application and scientific advancement of the Promoting Action on Research Implementation in Health Services (PARiHS) framework
Theoretical Domains Framework
In 2005, Michie and colleagues published the Theoretical Domains Framework in BMJ Quality & Safety , the result of a consensus process to develop a theoretical framework for implementation research. The primary goals of the development team were to determine key theoretical constructs for studying evidence-based practice implementation and for developing strategies for effective implementation, and for these constructs to be accessible and meaningful across disciplines.
The Theoretical Domains Framework (TDF) is an integrative framework developed to facilitate the investigation of determinants of behavior change and the design of behavior change interventions. Unlike a specific theory, the TDF does not propose testable relationships between elements; instead, it provides a theoretical lens through which to view the cognitive, affective, social, and environmental influences on behavior. Researchers use the TDF to assess implementation problems, design interventions, and understand change processes.
- ✪ Validation of the theoretical domains framework for use in behaviour change and implementation research ( Implementation Science , 2012)
- ✪ Theoretical domains framework to assess barriers to change for planning health care quality interventions: a systematic literature review ( Journal of Multidisciplinary Healthcare , 2016)
- ✪ Combined use of the Consolidated Framework for Implementation Research (CFIR) and the Theoretical Domains Framework (TDF): a systematic review ( Implementation Science , 2017)
- ✪ Applying the Theoretical Domains Framework to identify barriers and targeted interventions to enhance nurses’ use of electronic medication management systems in two Australian hospitals ( Implementation Science , 2017)
- ✪ A guide to using the Theoretical Domains Framework of behaviour change to investigate implementation problems ( Implementation Science , 2017)
- ✪ Hospitals Implementing Changes in Law to Protect Children of Ill Parents: A Cross-Sectional Study ( BMC Health Services Research , 2018)
- Addressing the Third Delay: Implementing a Novel Obstetric Triage System in Ghana ( BMJ Global Health , 2018)
Classic Theories
Behavioral theories.
In 2005 the NIH published ✪ Theory at a Glance: A Guide For Health Promotion Practice 2.0 , an overview of behavior change theories. Below are selected theories from the intrapersonal and interpersonal ecological levels most relevant to implementation science.
Intrapersonal Theories
There are two intrapersonal behavioral theories most often used to interpret individual behavior variation:
The Health Belief Model : An initial theory of health behavior, the HBM arose from work in the 1950s by a group of social psychologists in the U.S. wishing to understand why health improvement services were not being used. The HBM posited that in the health behavior context, readiness to act arises from six factors: perceived susceptibility , perceived severity . perceived benefits , perceived barriers , a cue to action , and self-efficacy . To learn more about the Health Belief Model, please read “Historical Origins of the Health Belief Model” ( Health Education Monographs ).
The Theory of Planned Behavior : This theory, developed by Ajzen in the late 1980s and formalized in 1991 , sees the primary driver of behavior as being behavioral intention . Through the lens of the TPB, behavioral intention is believed to be influenced by an individual’s attitude , their perception of peers’ subjective norms , and the individual’s perceived behavioral control .
Interpersonal Theories
At the interpersonal behavior level , where individual behavior is influenced by a social environment, Social Cognitive Theory is the most widely used theory in health behavior research.
Social Cognitive Theory : Published by Bandera in the 1978 article, Self-efficacy: Toward a unifying theory of behavioral change , SCT consists of six main constructs: reciprocal determinism , behavioral capability , expectations , observational learning , reinforcements , and self-efficacy (which is seen as the most important personal factor in changing behavior).
Examples of use in implementation science:
The Health Belief Model
- ✪ Using technology for improving population health: comparing classroom vs. online training for peer community health advisors in African American churches ( Implementation Science , 2015)
The Theory of Planned Behavior
- ✪ Assessing mental health clinicians’ intentions to adopt evidence-based treatments: reliability and validity testing of the evidence-based treatment intentions scale ( Implementation Science , 2016)
Social Cognitive Theory
- ✪ Systematic development of a theory-informed multifaceted behavioural intervention to increase physical activity of adults with type 2 diabetes in routine primary care: Movement as Medicine for Type 2 Diabetes ( Implementation Science , 2016)
Diffusion of Innovation
Diffusion of Innovation Theory is one of the oldest social science theories. It originated in communication to explain how, over time, an idea or product gains momentum and diffuses (or spreads) through a specific population or social system. The theory describes the pattern and speed at which new ideas, practices, or products spread through a population. Diffusion of Innovation theory has its roots in the early twentieth century, but the modern theory is credited to Everett Rogers with his publication in 1962 of Diffusion of Innovations .
This theory holds that adopters of an innovation can be split into five categories that distribute in a Bell curve over time: innovators (2.5%), early adopters (13.5%), early majority (34%), late majority (34%) and laggards (16%). Further, the theory states that any given adopter’s desire and ability to adopt an innovation is individual, based on information about, exposure to, and experience of the innovation and adoption process.
- Diffusion of preventive innovations ( Addictive Behaviors , 2002)
- ✪ Diffusion of Innovation Theory ( Canadian Journal of Nursing Informatics , 2011)
- Diffusion Of Innovations Theory, Principles, And Practice ( Health Affairs , 2018)
- 💻 Making Sense of Diffusion of Innovations Theory (University of Pittsburgh DISC)
- 💻 5 Common Misconceptions about Diffusion of Innovations Theory (University of Pittsburgh DISC)
Organizational Theory
Organizational theory plays a crucial role in implementation science, offering valuable insights into the complex interactions between organizations and their external environments. In 2017 Dr. Sarah Birken and colleagues published their application of four organizational theories to published accounts of evidence-based program implementation. The objective was to determine whether these theories could help explain implementation success by shedding light on the impact of the external environment on the implementing organizations.
Their paper, ✪ Organizational theory for dissemination and implementation research , published in the journal Implementation Science utilized transaction cost economics theory , institutional theory , contingency theories , and resource dependency theory for this work.
In 2019, Dr. Jennifer Leeman and colleagues applied these same three organizational theories to case studies of the implementation of colorectal cancer screening interventions in Federally Qualified Health Centers, in ✪ Advancing the use of organization theory in implementation science ( Preventive Medicine , 2019).
Organizational theory provides a lens through which implementation researchers can better comprehend the intricate relationships between organizations and their surroundings, ultimately enhancing the effectiveness of implementation efforts. Learn more in Leeman et al.’s 2022 article, Applying Theory to Explain the Influence of Factors External to an Organization on the Implementation of an Evidence-Based Intervention , and Birken et al.’s 2023 article, Toward a more comprehensive understanding of organizational influences on implementation: the organization theory for implementation science framework .
Implementation Theories
Implementation climate.
Implementation climate refers to a shared perception among intended users of an innovation within an organization. It reflects the extent to which an organization’s implementation policies and practices encourage, cultivate, and reward the use of that innovation. In other words, a strong implementation climate indicates that innovation use is expected, supported, and rewarded, leading to more consistent and high-quality implementation within the organization. This construct is particularly relevant for innovations that require coordinated behavior change by multiple organizational members for successful implementation and anticipated benefits. ✪ The meaning and measurement of implementation climate (2011) by Weiner, Belden, Bergmire, and Johnston, posits that the extent to which organizational members perceive that the use of an innovation is expected, supported, and rewarded, is positively associated with implementation effectiveness.
- ✪ A stepped-wedge randomized trial investigating the effect of the Leadership and Organizational Change for Implementation (LOCI) intervention on implementation and transformational leadership, and implementation climate ( BMC Health Services Research , 2022)
- ✪ Context matters: measuring implementation climate among individuals and groups ( Implementation Science , 2014)
- ✪ Determining the predictors of innovation implementation in healthcare: a quantitative analysis of implementation effectiveness ( BMC Health Services Research , 2015)
- ✪ Testing a theory of strategic implementation leadership, implementation climate, and clinicians’ use of evidence-based practice: a 5-year panel analysis ( Implementation Science , 2020)
- ✪ Individual-level associations between implementation leadership, climate, and anticipated outcomes: a time-lagged mediation analysis ( Implementation Science Communications , 2023)
Normalization Process Theory
Normalization Process Theory (NPT) is a sociological theory that helps us understand the dynamics of implementing, embedding, and integrating new technologies or complex interventions in healthcare. It identifies and explains key mechanisms that promote or inhibit the successful implementation of health techniques, technologies, and other interventions. Researchers and practitioners use NPT to rigorously study implementation processes, providing a conceptual vocabulary for analyzing the success or failure of specific projects. Essentially, NPT sheds light on how new practices become routinely embedded in everyday healthcare practice.
In 2010, Elizabeth Murray and colleagues published ✪ Normalization process theory: a framework for developing, evaluating and implementing complex interventions , comprised of four major components: Coherence, Cognitive Participation, Collective Action, and Reflexive Monitoring. The authors argued that using normalization process theory could enable researchers to think through issues of implementation concurrently while designing a complex intervention and its evaluation. They additionally held that normalization process theory could improve trial design by highlighting possible recruitment or data collection issues.
- ✪ Development of a theory of implementation and integration: Normalization Process Theory ( Implementation Science , 2009)
- ✪ Implementation, context and complexity ( Implementation Science , 2016)
- ✪ Understanding implementation context and social processes through integrating Normalization Process Theory (NPT) and the Consolidated Framework for Implementation Research (CFIR) ( Implementation Science Communications , 2022)
- ✪ Combining Realist approaches and Normalization Process Theory to understand implementation: a systematic review ( Implementation Science Communications , 2021)
- ✪ Using Normalization Process Theory in feasibility studies and process evaluations of complex healthcare interventions: a systematic review ( Implementation Science , 2018)
- ✪ Improving the normalization of complex interventions: part 1 – development of the NoMAD instrument for assessing implementation work based on normalization process theory (NPT) ( BMC Medical Research Methodology , 2018)
- 🎥 Strategic Intentions and Everyday Practices: What do Normalization Processes Look Like?
Organizational Readiness for Change
Organizational Readiness for Change is a multi-level, multi-faceted construct that plays a crucial role in successful implementation of complex changes in healthcare settings. At the organizational level, it refers to two key components: change commitment (organizational members’ shared resolve to implement a change) and change efficacy (their shared belief in their collective capability to carry out the change). This theory suggests that organizational readiness for change varies based on how much members value the change and how favorably they appraise factors like task demands, resource availability, and situational context. When readiness is high, members are more likely to initiate change, persist, and exhibit cooperative behavior, leading to more effective implementation.
In 2009, Bryan Weiner developed a theory of organizational readiness for change to address the lack of theoretical development or empirical study of the commonly used construct. In the Open Access (✪) development article, organizational readiness for change is conceptually defined and a theory of its determinants and outcomes is developed. The focus on the organizational level of analysis filled a theoretical gap necessary to address in order to refine approaches to improving healthcare delivery entailing collective behavior change and in 2014, Shea et al published a measure of organizational readiness for implementing change , based on Weiner’s 2009 theory, available Open Access (✪) in the journal Implementation Science .
- Review: Conceptualization and Measurement of Organizational Readiness for Change ( Medical Care Research and Review , 2008)
- ✪ Unpacking organizational readiness for change: an updated systematic review and content analysis of assessments ( BMC Health Services Research , 2020)
- ✪ Towards evidence-based palliative care in nursing homes in Sweden: a qualitative study informed by the organizational readiness to change theory ( Implementation Science , 2018)
- ✪ Assessing the reliability and validity of the Danish version of Organizational Readiness for Implementing Change (ORIC) ( Implementation Science , 2018)
- ✪ Psychometric properties of two implementation measures: Normalization MeAsure Development questionnaire (NoMAD) and organizational readiness for implementing change (ORIC) ( Implementation Science Research and Practice , 2024)
Evaluation Frameworks
Used to systematically evaluate implementation success, framework for reporting adaptations and modifications-enhanced (frame).
The FRAME (Framework for Reporting Adaptations and Modifications-Enhanced) is an expanded framework designed to characterize modifications made to evidence-based interventions during implementation. It was developed to address limitations in the original framework (Framework for Modification and Adaptations), which did not fully capture certain aspects of modification and adaptation. The updated FRAME includes the following eight components:
- Timing and Process: Describes when and how the modification occurred during implementation.
- Planned vs. Unplanned: Differentiates between planned/proactive adaptations and unplanned/reactive modifications.
- Decision-Maker: Identifies who determined that the modification should be made.
- Modified Element: Specifies what aspect of the intervention was modified.
- Level of Delivery: Indicates the level (e.g., individual, organization) at which the modification occurred.
- Context or Content-Level Modifications: Describes the type or nature of the modification.
- Fidelity Consistency: Assesses the extent to which the modification aligns with fidelity.
- Reasons for Modification: Includes both the intent/goal of the modification (e.g., cost reduction) and contextual factors that influenced the decision.
The FRAME can be used to support research on the timing, nature, goals, and impact of modifications to evidence-based interventions. Additionally, there is a related tool called FRAME-IS (Framework for Reporting Adaptations and Modifications to Implementation Strategies) , which focuses on documenting modifications to implementation strategies. Both tools aim to enhance our understanding of how adaptations and modifications influence implementation outcomes.
- Using a stakeholder-engaged, iterative, and systematic approach to adapting collaborative decision skills training for implementation in VA psychosocial rehabilitation and recovery centers ( BMC Health Services Research , 2022)
Implementation Outcomes Framework
In their 2011 publication, Proctor and colleagues proposed that implementation outcomes should be distinct from service outcomes or clinical outcomes. They identified eight discrete implementation outcomes and proposed a taxonomy to define them:
- Acceptability: The perception among implementation stakeholders that a given treatment, service, practice, or innovation is agreeable, palatable, or satisfactory
- Adoption: The intention, initial decision, or action to try or employ an innovation or evidence-based practice
- Appropriateness: The perceived fit, relevance, or compatibility of the innovation or evidence-based practice for a given practice setting, provider, or consumer; and/or perceived fit of the innovation to address a particular issue or problem
- Feasibility: The extent to which a new treatment, or an innovation, can be successfully used or carried out within a given agency or setting
- Fidelity: The degree to which an intervention was implemented as it was prescribed in the original protocol or as it was intended by the program developers
- Implementation cost: The cost impact of an implementation effort
- Penetration: The integration of a practice within a service setting and its subsystems
- S ustainability: The extent to which a newly implemented treatment is maintained or institutionalized within a service setting’s ongoing, stable operations
The framework development article, ✪ Outcomes for Implementation Research: Conceptual Distinctions, Measurement Challenges, and Research Agenda , is available through Administration and Policy in Mental Health and Mental Health Services Research . In 2023, Dr. Proctor and several colleagues published a follow up Ten years of implementation outcomes research: a scoping review in the journal Implementation Science , a scoping review of ‘the field’s progress in implementation outcomes research.’
- Toward Evidence-Based Measures of Implementation: Examining the Relationship Between Implementation Outcomes and Client Outcomes ( Journal of Substance Abuse Treatment , 2016)
- ✪ Toward criteria for pragmatic measurement in implementation research and practice: a stakeholder-driven approach using concept mapping ( Implementation Science , 2017)
- ✪ German language questionnaires for assessing implementation constructs and outcomes of psychosocial and health-related interventions: a systematic review ( Implementation Science , 2018)
- The Elusive Search for Success: Defining and Measuring Implementation Outcomes in a Real-World Hospital Trial ( Frontiers In Public Health , 2019)
RE-AIM (Reach, Efficacy, Adoption, Implementation, Maintenance)
The RE-AIM framework helps program planners, evaluators, and researchers consider five dimensions when designing, implementing, and assessing interventions:
- Reach: The extent to which an intervention reaches the intended target population, considering both the absolute number of participants and the representativeness of those participants
- Effectiveness: The impact of the intervention on relevant outcomes, assessing whether the intervention achieves its intended goals and produces positive results
- Adoption: The willingness of organizations or individuals to implement the intervention, considering factors such as organizational buy-in, acceptance, and readiness for change
- Implementation: How well the intervention is delivered in practice, looking at fidelity (adherence to the intervention components), quality, and consistency of delivery
- Maintenance: The long-term sustainability of the intervention, considering whether the program continues to be effective and is integrated into routine practice over time
In 1999, authors Glasgow, Vogt, and Boles developed this framework because they felt tightly controlled efficacy studies weren’t very helpful in informing program scale-up or in understanding actual public health impact of an intervention. The RE-AIM framework has been refined over time to guide the design and evaluation of complex interventions in order to maximize real-life public health impact.
This framework helps researchers collect information needed to translate research to effective practice, and may also be used to guide implementation and potential scale-up activities. You can read the original framework development article in The American Journal of Public Health . Additional resources, support, and publications on the RE-AIM framework can be found at RE-AIM.org . The 2021 special issue of Frontiers in Public Health titled Use of the RE-AIM Framework: Translating Research to Practice with Novel Applications and Emerging Directions includes more than 20 articles on RE-AIM.
- What Does It Mean to “Employ” the RE-AIM Model? ( Evaluation & the Health Professions , 2012)
- The RE-AIM Framework: A Systematic Review of Use Over Time (The American Journal of Public Health , 2013)
- ✪ Fidelity to and comparative results across behavioral interventions evaluated through the RE-AIM framework: a systematic review ( Systematic Reviews , 2015)
- ✪ Qualitative approaches to use of the RE-AIM framework: rationale and methods ( BMC Health Services Research , 2018)
- ✪ RE-AIM in Clinical, Community, and Corporate Settings: Perspectives, Strategies, and Recommendations to Enhance Public Health Impact ( Frontiers in Public Health , 2018)
- ✪ RE-AIM Planning and Evaluation Framework: Adapting to New Science and Practice With a 20-Year Review ( Frontiers in Public Health , 2019)
- ✪ RE-AIM in the Real World: Use of the RE-AIM Framework for Program Planning and Evaluation in Clinical and Community Settings ( Frontiers in Public Health , 2019)
- 💻 How to use RE-AIM (University of Pittsburgh DISC)
Stages of Implementation Completion (SIC)
Saldana’s Stages of Implementation Completion (SIC) is an eight-stage tool that assesses the implementation process and milestones across three phases: pre-implementation, implementation, and sustainability. It helps measure the duration (time to complete a stage), proportion (of stage activities completed), and overall progress of a site in the implementation process. The SIC aims to bridge the gap between the implementation process and associated costs. The eight stages of the SIC are:
- Engagement: Initial involvement and commitment to implementing the practice
- Consideration of Feasibility: Assessing whether the practice can be feasibly implemented
- Readiness Planning: Preparing for implementation by addressing organizational readiness
- Staff Hired and Trained: Recruiting and training staff for implementation
- Fidelity Monitoring Processes in Place: Establishing processes to monitor fidelity to the practice
- Services and Consultation Begin: Actual implementation of the practice
- Ongoing Services and Fidelity Monitoring: Continuation of services and fidelity monitoring
- Competency: Ensuring staff competence in delivering the practice
- ✪ The stages of implementation completion for evidence-based practice: Protocol for a mixed methods study . ( Implementation Science , 2014)
- ✪ Agency leaders’ assessments of feasibility and desirability of implementation of evidence-based practices in youth-serving organizations using the Stages of Implementation Completion ( Frontiers in Public Health , 2018)
- ✪ Economic evaluation in implementation science: Making the business case for implementation strategies ( Psychiatry Research , 2020)
- Scaling Implementation of Collaborative Care for Depression: Adaptation of the Stages of Implementation Completion (SIC) ( Administration and Policy in Mental Health and Mental Health Services Research , 2020)
- Adapting the stages of implementation completion to an evidence-based implementation strategy: The development of the NIATx stages of implementation completion ( Implementation Research and Practice , 2023)
PAUSE AND REFLECT
Does the T/M/F:
❯ specify the social, cultural, economic, and political contexts of the research?
❯ account for the needs and contexts of various demographic groups impacted, particularly those who are historically or currently marginalized or underserved?
❯ recognize and aim to dismantle existing power structures that contribute to inequities, including consideration of who has decision-making power and how it can be equitably distributed?
❯ emphasize building trusting relationships with communities and incorporating community-defined evidence so efforts are culturally relevant?
❯ address macro-, meso-, and micro-level influences on equity?
❯ encourage ongoing critical reflection on how well it advances equity and continue to identify areas for improvement?
- USC Libraries
- Research Guides
Organizing Your Social Sciences Research Paper
- The C.A.R.S. Model
- Purpose of Guide
- Design Flaws to Avoid
- Independent and Dependent Variables
- Glossary of Research Terms
- Reading Research Effectively
- Narrowing a Topic Idea
- Broadening a Topic Idea
- Extending the Timeliness of a Topic Idea
- Academic Writing Style
- Applying Critical Thinking
- Choosing a Title
- Making an Outline
- Paragraph Development
- Research Process Video Series
- Executive Summary
- Background Information
- The Research Problem/Question
- Theoretical Framework
- Citation Tracking
- Content Alert Services
- Evaluating Sources
- Primary Sources
- Secondary Sources
- Tiertiary Sources
- Scholarly vs. Popular Publications
- Qualitative Methods
- Quantitative Methods
- Insiderness
- Using Non-Textual Elements
- Limitations of the Study
- Common Grammar Mistakes
- Writing Concisely
- Avoiding Plagiarism
- Footnotes or Endnotes?
- Further Readings
- Generative AI and Writing
- USC Libraries Tutorials and Other Guides
- Bibliography
Introduction
The Creating a Research Space [C.A.R.S.] Model was developed by John Swales based upon his analysis of journal articles representing a variety of discipline-based writing practices.* His model attempts to explain and describe the organizational pattern of writing the introduction to scholarly research studies. Following the C.A.R.S. Model can be a useful approach because it can help you to: 1) begin the writing process [getting started is often the most difficult task]; 2) understand the way in which an introduction sets the stage for the rest of your paper; and, 3) assess how the introduction fits within the larger scope of your study. The model assumes that writers follow a general organizational pattern in response to two types of challenges [“competitions”] relating to establishing a presence within a particular domain of research: 1) the competition to create a rhetorical space and, 2) the competition to attract readers into that space. The model proposes three actions [Swales calls them “moves”], accompanied by specific steps, that reflect the development of an effective introduction for a research paper. These “moves” and steps can be used as a template for writing the introduction to your own social sciences research papers.
Despite the unnecessary jargon introduced by Swales, this approach can be useful in breaking down the essential elements of a paper's introduction, particularly if the research problem you are investigating is complex and multi-layered.
* Swales, John and Christine B. Feak. Academic Writing for Graduate Students: Essential Skills and Tasks . 3rd edition. Ann Arbor, MI: University of Michigan Press, 2012.
"Introductions." The Writing Lab and The OWL. Purdue University; Coffin, Caroline and Rupert Wegerif. “How to Write a Standard Research Article.” Inspiring Academic Practice at the University of Exeter; Kayfetz, Janet. "Academic Writing Workshop." University of California, Santa Barbara, Fall 2009; Pennington, Ken. "The Introduction Section: Creating a Research Space CARS Model." Language Centre, Helsinki University of Technology, 2005.
Creating a Research Space Move 1: Establishing a Territory [the situation] This is generally accomplished in two ways. First, by demonstrating that a general area of research is important, critical, interesting, problematic, relevant, or otherwise worthy of investigation and, second, by introducing and reviewing key sources of prior research in that area to show where gaps exist or where prior research has been inadequate in addressing the research problem. The steps taken to achieve this would be:
- Step 1 -- Claiming importance of, and/or [writing action = describing the research problem and providing evidence to support why the topic is important to study]
- Step 2 -- Making topic generalizations, and/or [writing action = providing statements about the current state of knowledge, consensus, practice or description of phenomena]
- Step 3 -- Reviewing items of previous research [writing action = synthesize prior research that further supports the need to study the research problem; this is not a literature review but more a reflection of key studies that have touched upon but perhaps not fully addressed the topic]
Move 2: Establishing a Niche [the problem] This action refers to making a clear and cogent argument that your particular area of research is important and possesses value. This can be done by indicating a specific gap in previous research, by challenging a broadly accepted assumption, by raising a question, a hypothesis, or need, or by extending previous knowledge in some way. The steps taken to achieve this would be:
- Step 1a -- Counter-claiming, or [writing action = introduce an opposing viewpoint or perspective or identify a gap in prior research that you believe has weakened or undermined the prevailing argument]
- Step 1b -- Indicating a gap, or [writing action = develop the research problem around a gap or understudied area of the literature]
- Step 1c -- Question-raising, or [writing action = similar to gap identification, this involves presenting key questions about the consequences of gaps in prior research that will be addressed by your study. For example, one could state, “Despite prior observations of voter behavior in local elections in Detroit, it remains unclear why do some single mothers choose to avoid....”]
- Step 1d -- Continuing a tradition [writing action = extend prior research to expand upon or clarify a research problem. This is often signaled with logical connecting terminology, such as, “hence,” “therefore,” “consequently,” “thus” or language that indicates a need. For example, one could state, “Consequently, these factors need to examined in more detail....” or “Evidence suggests an interesting correlation, therefore, it is desirable to survey different respondents....”]
Move 3: Occupying the Niche [the solution] The final "move" is to announce the means by which your study will contribute new knowledge or new understanding in contrast to prior research on the topic. This is also where you describe the remaining organizational structure of the paper. The steps taken to achieve this would be:
- Step 1a -- Outlining purposes, or [writing action = answering the “So What?” question. Explain in clear language the objectives of your study]
- Step 1b -- Announcing present research [writing action = describe the purpose of your study in terms of what the research is going to do or accomplish. In the social sciences, this also relates to answering the “So What?” question]
- Step 2 -- Announcing principle findings [writing action = present a brief, general summary of key findings written, such as, “The findings indicate a need for...,” or “The research suggests four approaches to....”]
- Step 3 -- Indicating article structure [writing action = state how the remainder of your paper is organized]
"Introductions." The Writing Lab and The OWL. Purdue University; Atai, Mahmood Reza. “Exploring Subdisciplinary Variations and Generic Structure of Applied Linguistics Research Article Introductions Using CARS Model.” The Journal of Applied Linguistics 2 (Fall 2009): 26-51; Chanel, Dana. "Research Article Introductions in Cultural Studies: A Genre Analysis Explorationn of Rhetorical Structure." The Journal of Teaching English for Specific and Academic Purposes 2 (2014): 1-20; Coffin, Caroline and Rupert Wegerif. “How to Write a Standard Research Article.” Inspiring Academic Practice at the University of Exeter; Kayfetz, Janet. "Academic Writing Workshop." University of California, Santa Barbara, Fall 2009; Pennington, Ken. "The Introduction Section: Creating a Research Space CARS Model." Language Centre, Helsinki University of Technology, 2005; Swales, John and Christine B. Feak. Academic Writing for Graduate Students: Essential Skills and Tasks . 3rd edition. Ann Arbor, MI: University of Michigan Press, 2012; Swales, John M. Genre Analysis: English in Academic and Research Settings . New York: Cambridge University Press, 1990; Chapter 5: Beginning Work. In Writing for Peer Reviewed Journals: Strategies for Getting Published . Pat Thomson and Barbara Kamler. (New York: Routledge, 2013), pp. 93-96.
Writing Tip
Swales showed that establishing a research niche [move 2] is often signaled by specific terminology that expresses a contrasting viewpoint, a critical evaluation of gaps in the literature, or a perceived weakness in prior research. The purpose of using these words is to draw a clear distinction between perceived deficiencies in previous studies and the research you are presenting that is intended to help resolve these deficiencies. Below is a table of common words used by authors.
NOTE: You may prefer not to adopt a negative stance in your writing when placing it within the context of prior research. In such cases, an alternative approach is to utilize a neutral, contrastive statement that expresses a new perspective without giving the appearance of trying to diminish the validity of other people's research. Examples of how to take a more neutral contrasting stance can be achieved in the following ways, with A representing the findings of prior research, B representing your research problem, and X representing one or more variables that have been investigated.
- Prior research has focused primarily on A , rather than on B ...
- Prior research into A can be beneficial but to rectify X , it is important to examine B ...
- These studies have placed an emphasis in the areas of A as opposed to describing B ...
- While prior studies have examined A , it may be preferable to contemplate the impact of B ...
- After consideration of A , it is important to also distinguish B ...
- The study of A has been thorough, but changing circumstances related to X support a need for examining [or revisiting] B ...
- Although research has been devoted to A , less attention has been paid to B ...
- Earlier research offers insights into the need for A , though consideration of B would be particularly helpful to address...
In each of these example statements, what follows the ellipsis is the justification for designing a study that approaches the problem in the way that contrasts with prior research, but which does not devalue its ongoing contributions to current knowledge and understanding.
Dretske, Fred I. “Contrastive Statements.” The Philosophical Review 81 (October 1972): 411-437; Kayfetz, Janet. "Academic Writing Workshop." University of California, Santa Barbara, Fall 2009; Pennington, Ken. "The Introduction Section: Creating a Research Space CARS Model." Language Centre, Helsinki University of Technology, 2005; Swales, John M. Genre Analysis: English in Academic and Research Settings . New York: Cambridge University Press, 1990
- << Previous: 4. The Introduction
- Next: Background Information >>
- Last Updated: Oct 24, 2024 10:02 AM
- URL: https://libguides.usc.edu/writingguide
- Search SF State Search SF State Button SF State This Site
SFSU researchers’ unique 3D maps might help solve the mysteries of octopus arms
Two new papers could help improve understanding of octopus arm function, development, evolution and more
Octopuses are fascinating. Their eight arms gracefully whip through water and can accomplish extraordinary tasks like using tools and opening jars. While humans have one spinal cord attached to their brain, in octopuses, it’s almost like each arm has its own spinal cord (minus the actual spine) and nervous system. These arms can even initiate a response without consulting the brain.
How octopus arms can do all this at a cellular level has largely remained a neuroscience mystery — one that’s proved difficult to study because of technological limitations and the expense of research. But now San Francisco State University researchers are starting to provide answers.
Trying to overcome those previous limitations, the San Francisco State researchers created three-dimensional molecular and anatomical maps of the inner neuronal circuitry of octopus arms. Their recent findings were published in two scientific papers in the journal Current Biology.
“Having [these two papers] converging at the same time means the amount we can learn from any single experiment is just astronomically higher,” SF State Biology Associate Department Chair and Assistant Professor Robyn Crook said of her lab’s research. “I would say these papers are really facilitating discovery in new ways.”
This research was supported by an Allen Distinguished Investigator Award, a Paul G. Allen Frontiers Group advised grant of the Paul G. Allen Family Foundation. Crook’s Allen Distinguished Investigator (ADI) grant was the first recipient in the California State University (CSU) system since the grant’s inception in 2010.
A traditional two-dimensional look at the octopus arm is comparable to taking a thin slice out of the middle of a fruit loaf. It’s difficult to know if distribution of fruits and nuts in that slice is representative of distribution and interactions throughout the loaf. Instead, postdoctoral fellow Gabrielle Winters-Bostwick and graduate student Diana Neacsu took multiple sections along the octopus arm to create 3D reconstructions of cell distribution and gross anatomy, respectively.
For her study , Winters-Bostwick used molecular tags to highlight different types of neurons. Seeing these neurons in a 3D reconstruction revealed that the cells at the tip of an octopus arm are different from those at the base closer to the central brain.
“This allows us to start hypothesizing and posing new questions thinking about how the cells communicate with one another,” she explained. “It’s basically building our arsenal and our toolkit to better understand the behavior and physiologies of octopuses.”
Using a different imaging approach (3D electron microscopy), Neacsu did a parallel project to create a 3D reconstruction mapping the structural organization of the components of the nervous system in the octopus arm. Her map revealed that there is symmetry in the organization of the ganglia and repeating patterns in nerve branching, blood vessels and more. Some of these patterns correspond to the octopus arm suckers, which are organized in a hexagonal lattice like rows of honeycomb. This repeating pattern is something they couldn’t see with just two suckers, Crook explained, highlighting the necessity of the 3D reconstruction of a large tissue.
“To see how closely the [nervous system structures] associated with the suckers was really surprising,” Neacsu said. “But it makes sense because the suckers play such a huge role in the octopus’s ecological niche, helping them hunt, sense and more.”
Crook is proud to say her team was able to do much of these projects in-house at SF State. Of particular importance was the recently acquired microscope (Leica STELLARIS) in the University’s on-campus Cellular and Molecular Imaging Center (CMIC) , which has trained over 1,000 students. “There are a lot of [universities] that don’t have a microscope like this. For us to have one here to do this work is kind of mind blowing,” Crook said. “[Winters-Bostwick’s] paper would not exist without that microscope.”
One of the major limiting factors in research — particularly cutting-edge projects like Crook’s — is the high price tag of equipment and computational tools. “The ADI grant has been transformative to have funds to do things in my lab that I would not have been able to do and to engage students on a really big scale,” Crook notes. “It’s been transformative for me as a PI but also for the students in my lab.”
The ADI project and Crook’s mentorship were instrumental for Neacsu, now a Ph.D. student at Katholieke Universiteit (KU) Leuven in Belgium. During her two years in Crook’s lab, Neacsu gained advanced technical skills and networked and collaborated with more senior researchers, and now she has more scientific research papers in the pipeline.
“Before I met her, I never really understood the concept of mentorship,” Neacsu said of Crook. “I kind of just thought [mentors] were teachers that are available during office hours.”
Neacsu’s and Winters-Bostwick’s papers enabled a myriad of research opportunities both within Crook’s lab and beyond. Other labs have already showed interest in using these tools for cephalopod neuroscience research.
The SF State team is looking at live tissues and seeing how they respond to chemical and mechanical stimulation, trying to understand neurons firing in real time. With the new 3D maps, they can make realistic predictions about what’s happening inside an octopus arm to create these responses. There are also a lot of evolutionary questions Crook’s lab is eager to answer.
“Why do you have an animal with this much complexity that doesn’t seem to follow the same rules as our other example — humans — of a very complex nervous system?” Crook asked. “There’s a lot of hypotheses. It might be functional. There might be something fundamentally different in the tasks octopus arms have to do. But it could also be an evolutionary accident.”
Learn more about research in SF State’s Department of Biology.
- College of Science and Engineering
Related News
- Email: [email protected]
- Telephone: (415) 338-1665
Office Hours
Quick links.
- SF State Magazine
- President's Messages
- University Calendar
- SF State in the Media
IMAGES
VIDEO
COMMENTS
A presentation on the definition, types, and application of research models and methodologies in information systems. Learn about the concepts, paradigms, theories, methods, and domains of research in this discipline.
Learn how to design a research strategy for answering your research question using empirical data. Compare different types of research design, such as qualitative, quantitative, experimental, and mixed-methods, and see examples of each.
Learn how to choose and use research methods for collecting and analyzing data. Compare qualitative and quantitative, primary and secondary, descriptive and experimental methods with examples and pros and cons.
What Are Theories. The terms theory and model have been defined in numerous ways, and there are at least as many ideas on how theories and models relate to each other (Bailer-Jones, Citation 2009).I understand theories as bodies of knowledge that are broad in scope and aim to explain robust phenomena.Models, on the other hand, are instantiations of theories, narrower in scope and often more ...
Learn what research paradigms are and how they shape your research methodology. Explore the four types of research paradigms: positivist, interpretivist, critical theory, and constructivist, and how to choose the right one for your study.
About Research Methods. This guide provides an overview of research methods, how to choose and use them, and supports and resources at UC Berkeley. As Patten and Newhart note in the book Understanding Research Methods, "Research methods are the building blocks of the scientific enterprise. They are the "how" for building systematic knowledge.
Learn about different types of research strategies and models, such as qualitative, quantitative, physical, theoretical, mathematical, mechanical and symbolic interactionist. Explore the advantages and disadvantages of modelling in various disciplines and contexts.
Developing a conceptual framework in research. Step 1: Choose your research question. Step 2: Select your independent and dependent variables. Step 3: Visualize your cause-and-effect relationship. Step 4: Identify other influencing variables. Frequently asked questions about conceptual models.
Learn how to conduct research that is both critical and inclusive. Doing Research Online. Learn to design and conduct online and digital research with videos, case studies, practice data and how-to guides. Foundations. Read bite-sized introductions to hundreds of research concepts and methods written by global experts. Medicine and Health.
Theoretical frameworks provide a particular perspective, or lens, through which to examine a topic. They come from many areas of academic research, resulting in psychological theories, social theories, organizational theories and economic theories among others. Theoretical frameworks are often used to define concepts and explain phenomena.
Abstract. This chapter discusses the research-driven development of models, theories and approaches for design and development. It begins by clarifying the types of models, theories and approaches considered. Desirable characteristics for each specific type are then outlined, and research methods for developing and evaluating them are discussed.
Process models: These describe or guide the translation of research evidence into practice, outlining the steps involved in implementing evidence-based practices. Determinants frameworks: These focus on understanding and explaining the factors that influence implementation outcomes, highlighting barriers and enablers (but may lack specific practical guidance).
Here, the author's aim is to integrate some of the core points and criticism raised, and provide a brief primer on theory formation, structured into three sections: (1) what are theories; (2) what are theories for; (3) and what are theories about. This is followed by a section dedicated to the question (4) how to develop theories.
The Creating a Research Space [C.A.R.S.] Model was developed by John Swales based upon his analysis of journal articles representing a variety of discipline-based writing practices.* His model attempts to explain and describe the organizational pattern of writing the introduction to scholarly research studies. Following the C.A.R.S. Model can ...
Learn how to choose the right type of research for your project based on the research aims, data, sampling, timescale, and location. Compare different types of research with examples and explanations.
exercise in statistical model fitting, and falls short of theory. building and testing in three ways. First, theories are absent, which fosters conflating statistical models with theoretical ...
This book chapter introduces a model of qualitative research design that treats design as a real entity, not simply a plan or protocol. It contrasts this model with traditional, linear approaches and illustrates its features with examples.
This research was supported by an Allen Distinguished Investigator Award, a Paul G. Allen Frontiers Group advised grant of the Paul G. Allen Family Foundation. Crook's Allen Distinguished Investigator (ADI) grant was the first recipient in the California State University (CSU) system since the grant's inception in 2010.
Action research models. Action research is often reflected in 3 action research models: operational (sometimes called technical), collaboration, and critical reflection. Operational (or technical) action research is usually visualized like a spiral following a series of steps, such as "planning → acting → observing → reflecting."
Revised on September 5, 2024. Qualitative research involves collecting and analyzing non-numerical data (e.g., text, video, or audio) to understand concepts, opinions, or experiences. It can be used to gather in-depth insights into a problem or generate new ideas for research. Qualitative research is the opposite of quantitative research, which ...