• Bipolar Disorder
  • Therapy Center
  • When To See a Therapist
  • Types of Therapy
  • Best Online Therapy
  • Best Couples Therapy
  • Best Family Therapy
  • Managing Stress
  • Sleep and Dreaming
  • Understanding Emotions
  • Self-Improvement
  • Healthy Relationships
  • Student Resources
  • Personality Types
  • Guided Meditations
  • Verywell Mind Insights
  • 2024 Verywell Mind 25
  • Mental Health in the Classroom
  • Editorial Process
  • Meet Our Review Board
  • Crisis Support

The Sapir-Whorf Hypothesis: How Language Influences How We Express Ourselves

Rachael is a New York-based writer and freelance writer for Verywell Mind, where she leverages her decades of personal experience with and research on mental illness—particularly ADHD and depression—to help readers better understand how their mind works and how to manage their mental health.

hypothesis definition in linguistics

Thomas Barwick / Getty Images

What to Know About the Sapir-Whorf Hypothesis

Real-world examples of linguistic relativity, linguistic relativity in psychology.

The Sapir-Whorf Hypothesis, also known as linguistic relativity, refers to the idea that the language a person speaks can influence their worldview, thought, and even how they experience and understand the world.

While more extreme versions of the hypothesis have largely been discredited, a growing body of research has demonstrated that language can meaningfully shape how we understand the world around us and even ourselves.

Keep reading to learn more about linguistic relativity, including some real-world examples of how it shapes thoughts, emotions, and behavior.  

The hypothesis is named after anthropologist and linguist Edward Sapir and his student, Benjamin Lee Whorf. While the hypothesis is named after them both, the two never actually formally co-authored a coherent hypothesis together.

This Hypothesis Aims to Figure Out How Language and Culture Are Connected

Sapir was interested in charting the difference in language and cultural worldviews, including how language and culture influence each other. Whorf took this work on how language and culture shape each other a step further to explore how different languages might shape thought and behavior.

Since then, the concept has evolved into multiple variations, some more credible than others.

Linguistic Determinism Is an Extreme Version of the Hypothesis

Linguistic determinism, for example, is a more extreme version suggesting that a person’s perception and thought are limited to the language they speak. An early example of linguistic determinism comes from Whorf himself who argued that the Hopi people in Arizona don’t conjugate verbs into past, present, and future tenses as English speakers do and that their words for units of time (like “day” or “hour”) were verbs rather than nouns.

From this, he concluded that the Hopi don’t view time as a physical object that can be counted out in minutes and hours the way English speakers do. Instead, Whorf argued, the Hopi view time as a formless process.

This was then taken by others to mean that the Hopi don’t have any concept of time—an extreme view that has since been repeatedly disproven.

There is some evidence for a more nuanced version of linguistic relativity, which suggests that the structure and vocabulary of the language you speak can influence how you understand the world around you. To understand this better, it helps to look at real-world examples of the effects language can have on thought and behavior.

Different Languages Express Colors Differently

Color is one of the most common examples of linguistic relativity. Most known languages have somewhere between two and twelve color terms, and the way colors are categorized varies widely. In English, for example, there are distinct categories for blue and green .

Blue and Green

But in Korean, there is one word that encompasses both. This doesn’t mean Korean speakers can’t see blue, it just means blue is understood as a variant of green rather than a distinct color category all its own.

In Russian, meanwhile, the colors that English speakers would lump under the umbrella term of “blue” are further subdivided into two distinct color categories, “siniy” and “goluboy.” They roughly correspond to light blue and dark blue in English. But to Russian speakers, they are as distinct as orange and brown .

In one study comparing English and Russian speakers, participants were shown a color square and then asked to choose which of the two color squares below it was the closest in shade to the first square.

The test specifically focused on varying shades of blue ranging from “siniy” to “goluboy.” Russian speakers were not only faster at selecting the matching color square but were more accurate in their selections.

The Way Location Is Expressed Varies Across Languages

This same variation occurs in other areas of language. For example, in Guugu Ymithirr, a language spoken by Aboriginal Australians, spatial orientation is always described in absolute terms of cardinal directions. While an English speaker would say the laptop is “in front of” you, a Guugu Ymithirr speaker would say it was north, south, west, or east of you.

As a result, Aboriginal Australians have to be constantly attuned to cardinal directions because their language requires it (just as Russian speakers develop a more instinctive ability to discern between shades of what English speakers call blue because their language requires it).

So when you ask a Guugu Ymithirr speaker to tell you which way south is, they can point in the right direction without a moment’s hesitation. Meanwhile, most English speakers would struggle to accurately identify South without the help of a compass or taking a moment to recall grade school lessons about how to find it.

The concept of these cardinal directions exists in English, but English speakers aren’t required to think about or use them on a daily basis so it’s not as intuitive or ingrained in how they orient themselves in space.

Just as with other aspects of thought and perception, the vocabulary and grammatical structure we have for thinking about or talking about what we feel doesn’t create our feelings, but it does shape how we understand them and, to an extent, how we experience them.

Words Help Us Put a Name to Our Emotions

For example, the ability to detect displeasure from a person’s face is universal. But in a language that has the words “angry” and “sad,” you can further distinguish what kind of displeasure you observe in their facial expression. This doesn’t mean humans never experienced anger or sadness before words for them emerged. But they may have struggled to understand or explain the subtle differences between different dimensions of displeasure.

In one study of English speakers, toddlers were shown a picture of a person with an angry facial expression. Then, they were given a set of pictures of people displaying different expressions including happy, sad, surprised, scared, disgusted, or angry. Researchers asked them to put all the pictures that matched the first angry face picture into a box.

The two-year-olds in the experiment tended to place all faces except happy faces into the box. But four-year-olds were more selective, often leaving out sad or fearful faces as well as happy faces. This suggests that as our vocabulary for talking about emotions expands, so does our ability to understand and distinguish those emotions.

But some research suggests the influence is not limited to just developing a wider vocabulary for categorizing emotions. Language may “also help constitute emotion by cohering sensations into specific perceptions of ‘anger,’ ‘disgust,’ ‘fear,’ etc.,” said Dr. Harold Hong, a board-certified psychiatrist at New Waters Recovery in North Carolina.

As our vocabulary for talking about emotions expands, so does our ability to understand and distinguish those emotions.

Words for emotions, like words for colors, are an attempt to categorize a spectrum of sensations into a handful of distinct categories. And, like color, there’s no objective or hard rule on where the boundaries between emotions should be which can lead to variation across languages in how emotions are categorized.

Emotions Are Categorized Differently in Different Languages

Just as different languages categorize color a little differently, researchers have also found differences in how emotions are categorized. In German, for example, there’s an emotion called “gemütlichkeit.”

While it’s usually translated as “cozy” or “ friendly ” in English, there really isn’t a direct translation. It refers to a particular kind of peace and sense of belonging that a person feels when surrounded by the people they love or feel connected to in a place they feel comfortable and free to be who they are.

Harold Hong, MD, Psychiatrist

The lack of a word for an emotion in a language does not mean that its speakers don't experience that emotion.

You may have felt gemütlichkeit when staying up with your friends to joke and play games at a sleepover. You may feel it when you visit home for the holidays and spend your time eating, laughing, and reminiscing with your family in the house you grew up in.

In Japanese, the word “amae” is just as difficult to translate into English. Usually, it’s translated as "spoiled child" or "presumed indulgence," as in making a request and assuming it will be indulged. But both of those have strong negative connotations in English and amae is a positive emotion .

Instead of being spoiled or coddled, it’s referring to that particular kind of trust and assurance that comes with being nurtured by someone and knowing that you can ask for what you want without worrying whether the other person might feel resentful or burdened by your request.

You might have felt amae when your car broke down and you immediately called your mom to pick you up, without having to worry for even a second whether or not she would drop everything to help you.

Regardless of which languages you speak, though, you’re capable of feeling both of these emotions. “The lack of a word for an emotion in a language does not mean that its speakers don't experience that emotion,” Dr. Hong explained.

What This Means For You

“While having the words to describe emotions can help us better understand and regulate them, it is possible to experience and express those emotions without specific labels for them.” Without the words for these feelings, you can still feel them but you just might not be able to identify them as readily or clearly as someone who does have those words. 

Rhee S. Lexicalization patterns in color naming in Korean . In: Raffaelli I, Katunar D, Kerovec B, eds. Studies in Functional and Structural Linguistics. Vol 78. John Benjamins Publishing Company; 2019:109-128. Doi:10.1075/sfsl.78.06rhe

Winawer J, Witthoft N, Frank MC, Wu L, Wade AR, Boroditsky L. Russian blues reveal effects of language on color discrimination . Proc Natl Acad Sci USA. 2007;104(19):7780-7785.  10.1073/pnas.0701644104

Lindquist KA, MacCormack JK, Shablack H. The role of language in emotion: predictions from psychological constructionism . Front Psychol. 2015;6. Doi:10.3389/fpsyg.2015.00444

By Rachael Green Rachael is a New York-based writer and freelance writer for Verywell Mind, where she leverages her decades of personal experience with and research on mental illness—particularly ADHD and depression—to help readers better understand how their mind works and how to manage their mental health.

Sapir–Whorf hypothesis (Linguistic Relativity Hypothesis)

Mia Belle Frothingham

Author, Researcher, Science Communicator

BA with minors in Psychology and Biology, MRes University of Edinburgh

Mia Belle Frothingham is a Harvard University graduate with a Bachelor of Arts in Sciences with minors in biology and psychology

Learn about our Editorial Process

Saul Mcleod, PhD

Editor-in-Chief for Simply Psychology

BSc (Hons) Psychology, MRes, PhD, University of Manchester

Saul Mcleod, PhD., is a qualified psychology teacher with over 18 years of experience in further and higher education. He has been published in peer-reviewed journals, including the Journal of Clinical Psychology.

Olivia Guy-Evans, MSc

Associate Editor for Simply Psychology

BSc (Hons) Psychology, MSc Psychology of Education

Olivia Guy-Evans is a writer and associate editor for Simply Psychology. She has previously worked in healthcare and educational sectors.

On This Page:

There are about seven thousand languages heard around the world – they all have different sounds, vocabularies, and structures. As you know, language plays a significant role in our lives.

But one intriguing question is – can it actually affect how we think?

Collection of talking people. Men and women with speech bubbles. Communication and interaction. Friends, students or colleagues. Cartoon flat vector illustrations isolated on white background

It is widely thought that reality and how one perceives the world is expressed in spoken words and are precisely the same as reality.

That is, perception and expression are understood to be synonymous, and it is assumed that speech is based on thoughts. This idea believes that what one says depends on how the world is encoded and decoded in the mind.

However, many believe the opposite.

In that, what one perceives is dependent on the spoken word. Basically, that thought depends on language, not the other way around.

What Is The Sapir-Whorf Hypothesis?

Twentieth-century linguists Edward Sapir and Benjamin Lee Whorf are known for this very principle and its popularization. Their joint theory, known as the Sapir-Whorf Hypothesis or, more commonly, the Theory of Linguistic Relativity, holds great significance in all scopes of communication theories.

The Sapir-Whorf hypothesis states that the grammatical and verbal structure of a person’s language influences how they perceive the world. It emphasizes that language either determines or influences one’s thoughts.

The Sapir-Whorf hypothesis states that people experience the world based on the structure of their language, and that linguistic categories shape and limit cognitive processes. It proposes that differences in language affect thought, perception, and behavior, so speakers of different languages think and act differently.

For example, different words mean various things in other languages. Not every word in all languages has an exact one-to-one translation in a foreign language.

Because of these small but crucial differences, using the wrong word within a particular language can have significant consequences.

The Sapir-Whorf hypothesis is sometimes called “linguistic relativity” or the “principle of linguistic relativity.” So while they have slightly different names, they refer to the same basic proposal about the relationship between language and thought.

How Language Influences Culture

Culture is defined by the values, norms, and beliefs of a society. Our culture can be considered a lens through which we undergo the world and develop a shared meaning of what occurs around us.

The language that we create and use is in response to the cultural and societal needs that arose. In other words, there is an apparent relationship between how we talk and how we perceive the world.

One crucial question that many intellectuals have asked is how our society’s language influences its culture.

Linguist and anthropologist Edward Sapir and his then-student Benjamin Whorf were interested in answering this question.

Together, they created the Sapir-Whorf hypothesis, which states that our thought processes predominantly determine how we look at the world.

Our language restricts our thought processes – our language shapes our reality. Simply, the language that we use shapes the way we think and how we see the world.

Since the Sapir-Whorf hypothesis theorizes that our language use shapes our perspective of the world, people who speak different languages have different views of the world.

In the 1920s, Benjamin Whorf was a Yale University graduate student studying with linguist Edward Sapir, who was considered the father of American linguistic anthropology.

Sapir was responsible for documenting and recording the cultures and languages of many Native American tribes disappearing at an alarming rate. He and his predecessors were well aware of the close relationship between language and culture.

Anthropologists like Sapir need to learn the language of the culture they are studying to understand the worldview of its speakers truly. Whorf believed that the opposite is also true, that language affects culture by influencing how its speakers think.

His hypothesis proposed that the words and structures of a language influence how its speaker behaves and feels about the world and, ultimately, the culture itself.

Simply put, Whorf believed that you see the world differently from another person who speaks another language due to the specific language you speak.

Human beings do not live in the matter-of-fact world alone, nor solitary in the world of social action as traditionally understood, but are very much at the pardon of the certain language which has become the medium of communication and expression for their society.

To a large extent, the real world is unconsciously built on habits in regard to the language of the group. We hear and see and otherwise experience broadly as we do because the language habits of our community predispose choices of interpretation.

Studies & Examples

The lexicon, or vocabulary, is the inventory of the articles a culture speaks about and has classified to understand the world around them and deal with it effectively.

For example, our modern life is dictated for many by the need to travel by some vehicle – cars, buses, trucks, SUVs, trains, etc. We, therefore, have thousands of words to talk about and mention, including types of models, vehicles, parts, or brands.

The most influential aspects of each culture are similarly reflected in the dictionary of its language. Among the societies living on the islands in the Pacific, fish have significant economic and cultural importance.

Therefore, this is reflected in the rich vocabulary that describes all aspects of the fish and the environments that islanders depend on for survival.

For example, there are over 1,000 fish species in Palau, and Palauan fishers knew, even long before biologists existed, details about the anatomy, behavior, growth patterns, and habitat of most of them – far more than modern biologists know today.

Whorf’s studies at Yale involved working with many Native American languages, including Hopi. He discovered that the Hopi language is quite different from English in many ways, especially regarding time.

Western cultures and languages view times as a flowing river that carries us continuously through the present, away from the past, and to the future.

Our grammar and system of verbs reflect this concept with particular tenses for past, present, and future.

We perceive this concept of time as universal in that all humans see it in the same way.

Although a speaker of Hopi has very different ideas, their language’s structure both reflects and shapes the way they think about time. Seemingly, the Hopi language has no present, past, or future tense; instead, they divide the world into manifested and unmanifest domains.

The manifested domain consists of the physical universe, including the present, the immediate past, and the future; the unmanifest domain consists of the remote past and the future and the world of dreams, thoughts, desires, and life forces.

Also, there are no words for minutes, minutes, or days of the week. Native Hopi speakers often had great difficulty adapting to life in the English-speaking world when it came to being on time for their job or other affairs.

It is due to the simple fact that this was not how they had been conditioned to behave concerning time in their Hopi world, which followed the phases of the moon and the movements of the sun.

Today, it is widely believed that some aspects of perception are affected by language.

One big problem with the original Sapir-Whorf hypothesis derives from the idea that if a person’s language has no word for a specific concept, then that person would not understand that concept.

Honestly, the idea that a mother tongue can restrict one’s understanding has been largely unaccepted. For example, in German, there is a term that means to take pleasure in another person’s unhappiness.

While there is no translatable equivalent in English, it just would not be accurate to say that English speakers have never experienced or would not be able to comprehend this emotion.

Just because there is no word for this in the English language does not mean English speakers are less equipped to feel or experience the meaning of the word.

Not to mention a “chicken and egg” problem with the theory.

Of course, languages are human creations, very much tools we invented and honed to suit our needs. Merely showing that speakers of diverse languages think differently does not tell us whether it is the language that shapes belief or the other way around.

Supporting Evidence

On the other hand, there is hard evidence that the language-associated habits we acquire play a role in how we view the world. And indeed, this is especially true for languages that attach genders to inanimate objects.

There was a study done that looked at how German and Spanish speakers view different things based on their given gender association in each respective language.

The results demonstrated that in describing things that are referred to as masculine in Spanish, speakers of the language marked them as having more male characteristics like “strong” and “long.” Similarly, these same items, which use feminine phrasings in German, were noted by German speakers as effeminate, like “beautiful” and “elegant.”

The findings imply that speakers of each language have developed preconceived notions of something being feminine or masculine, not due to the objects” characteristics or appearances but because of how they are categorized in their native language.

It is important to remember that the Theory of Linguistic Relativity (Sapir-Whorf Hypothesis) also successfully achieves openness. The theory is shown as a window where we view the cognitive process, not as an absolute.

It is set forth to look at a phenomenon differently than one usually would. Furthermore, the Sapir-Whorf Hypothesis is very simple and logically sound. Understandably, one’s atmosphere and culture will affect decoding.

Likewise, in studies done by the authors of the theory, many Native American tribes do not have a word for particular things because they do not exist in their lives. The logical simplism of this idea of relativism provides parsimony.

Truly, the Sapir-Whorf Hypothesis makes sense. It can be utilized in describing great numerous misunderstandings in everyday life. When a Pennsylvanian says “yuns,” it does not make any sense to a Californian, but when examined, it is just another word for “you all.”

The Linguistic Relativity Theory addresses this and suggests that it is all relative. This concept of relativity passes outside dialect boundaries and delves into the world of language – from different countries and, consequently, from mind to mind.

Is language reality honestly because of thought, or is it thought which occurs because of language? The Sapir-Whorf Hypothesis very transparently presents a view of reality being expressed in language and thus forming in thought.

The principles rehashed in it show a reasonable and even simple idea of how one perceives the world, but the question is still arguable: thought then language or language then thought?

Modern Relevance

Regardless of its age, the Sapir-Whorf hypothesis, or the Linguistic Relativity Theory, has continued to force itself into linguistic conversations, even including pop culture.

The idea was just recently revisited in the movie “Arrival,” – a science fiction film that engagingly explores the ways in which an alien language can affect and alter human thinking.

And even if some of the most drastic claims of the theory have been debunked or argued against, the idea has continued its relevance, and that does say something about its importance.

Hypotheses, thoughts, and intellectual musings do not need to be totally accurate to remain in the public eye as long as they make us think and question the world – and the Sapir-Whorf Hypothesis does precisely that.

The theory does not only make us question linguistic theory and our own language but also our very existence and how our perceptions might shape what exists in this world.

There are generalities that we can expect every person to encounter in their day-to-day life – in relationships, love, work, sadness, and so on. But thinking about the more granular disparities experienced by those in diverse circumstances, linguistic or otherwise, helps us realize that there is more to the story than ours.

And beautifully, at the same time, the Sapir-Whorf Hypothesis reiterates the fact that we are more alike than we are different, regardless of the language we speak.

Isn’t it just amazing that linguistic diversity just reveals to us how ingenious and flexible the human mind is – human minds have invented not one cognitive universe but, indeed, seven thousand!

Kay, P., & Kempton, W. (1984). What is the Sapir‐Whorf hypothesis?. American anthropologist, 86(1), 65-79.

Whorf, B. L. (1952). Language, mind, and reality. ETC: A review of general semantics, 167-188.

Whorf, B. L. (1997). The relation of habitual thought and behavior to language. In Sociolinguistics (pp. 443-463). Palgrave, London.

Whorf, B. L. (2012). Language, thought, and reality: Selected writings of Benjamin Lee Whorf. MIT press.

Print Friendly, PDF & Email

Related Articles

Automatic Processing in Psychology: Definition & Examples

Cognitive Psychology

Automatic Processing in Psychology: Definition & Examples

Controlled Processing in Psychology: Definition & Examples

Controlled Processing in Psychology: Definition & Examples

How Ego Depletion Can Drain Your Willpower

How Ego Depletion Can Drain Your Willpower

What is the Default Mode Network?

What is the Default Mode Network?

Theories of Selective Attention in Psychology

Availability Heuristic and Decision Making

Availability Heuristic and Decision Making

The Sapir-Whorf Hypothesis Linguistic Theory

DrAfter123/Getty Images

  • An Introduction to Punctuation
  • Ph.D., Rhetoric and English, University of Georgia
  • M.A., Modern English and American Literature, University of Leicester
  • B.A., English, State University of New York

The Sapir-Whorf hypothesis is the  linguistic theory that the semantic structure of a language shapes or limits the ways in which a speaker forms conceptions of the world. It came about in 1929. The theory is named after the American anthropological linguist Edward Sapir (1884–1939) and his student Benjamin Whorf (1897–1941). It is also known as the   theory of linguistic relativity, linguistic relativism, linguistic determinism, Whorfian hypothesis , and Whorfianism .

History of the Theory

The idea that a person's native language determines how he or she thinks was popular among behaviorists of the 1930s and on until cognitive psychology theories came about, beginning in the 1950s and increasing in influence in the 1960s. (Behaviorism taught that behavior is a result of external conditioning and doesn't take feelings, emotions, and thoughts into account as affecting behavior. Cognitive psychology studies mental processes such as creative thinking, problem-solving, and attention.)

Author Lera Boroditsky gave some background on ideas about the connections between languages and thought:

"The question of whether languages shape the way we think goes back centuries; Charlemagne proclaimed that 'to have a second language is to have a second soul.' But the idea went out of favor with scientists when  Noam Chomsky 's theories of language gained popularity in the 1960s and '70s. Dr. Chomsky proposed that there is a  universal grammar  for all human languages—essentially, that languages don't really differ from one another in significant ways...." ("Lost in Translation." "The Wall Street Journal," July 30, 2010)

The Sapir-Whorf hypothesis was taught in courses through the early 1970s and had become widely accepted as truth, but then it fell out of favor. By the 1990s, the Sapir-Whorf hypothesis was left for dead, author Steven Pinker wrote. "The cognitive revolution in psychology, which made the study of pure thought possible, and a number of studies showing meager effects of language on concepts, appeared to kill the concept in the 1990s... But recently it has been resurrected, and 'neo-Whorfianism' is now an active research topic in  psycholinguistics ." ("The Stuff of Thought. "Viking, 2007)

Neo-Whorfianism is essentially a weaker version of the Sapir-Whorf hypothesis and says that language  influences  a speaker's view of the world but does not inescapably determine it.

The Theory's Flaws

One big problem with the original Sapir-Whorf hypothesis stems from the idea that if a person's language has no word for a particular concept, then that person would not be able to understand that concept, which is untrue. Language doesn't necessarily control humans' ability to reason or have an emotional response to something or some idea. For example, take the German word  sturmfrei , which essentially is the feeling when you have the whole house to yourself because your parents or roommates are away. Just because English doesn't have a single word for the idea doesn't mean that Americans can't understand the concept.

There's also the "chicken and egg" problem with the theory. "Languages, of course, are human creations, tools we invent and hone to suit our needs," Boroditsky continued. "Simply showing that speakers of different languages think differently doesn't tell us whether it's language that shapes thought or the other way around."

  • Definition and Discussion of Chomskyan Linguistics
  • Cognitive Grammar
  • Generative Grammar: Definition and Examples
  • Universal Grammar (UG)
  • Transformational Grammar (TG) Definition and Examples
  • Linguistic Performance
  • What Is a Natural Language?
  • The Theory of Poverty of the Stimulus in Language Development
  • Linguistic Competence: Definition and Examples
  • The Definition and Usage of Optimality Theory
  • What Is Linguistic Functionalism?
  • 24 Words Worth Borrowing From Other Languages
  • Definition and Examples of Case Grammar
  • Biography of Noam Chomsky, Writer and Father of Modern Linguistics
  • An Introduction to Semantics
  • What Is Relevance Theory in Terms of Communication?

Whorfian Hypothesis

  • Reference work entry
  • pp 1566–1567
  • Cite this reference work entry

hypothesis definition in linguistics

  • Seongwon Yun 3 &
  • Shelia M. Kennison 3  

749 Accesses

Linguistic determinism ; Linguistic relativity ; Sapir-Whorf hypothesis

The hypothesis suggests that human thought is influenced by the language one speaks.

Description

The term Whorfian Hypothesis takes its name from Benjamin Lee Whorf (1876–1941) who claimed that the language one speaks influences one’s thinking [ 7 ]. Whorf was an amateur linguist who studied with the anthropologist Edward Sapir in the 1920s and 1930s. The term Sapir-Whorf Hypothesis is also used to refer to their view that language determines thinking. Linguistic determinism and linguistic relativity are also terms referring to the notion that the characteristics of one’s language shape one’s cognition.

The hypothesis was developed following observations of cross language differences in linguistic structure and speculations about how such differences might impact speakers’ thinking. For example, Whorf noted that in the Hopi language, there was little to no grammatical marking for tense. Utterances...

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
  • Available as EPUB and PDF
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Berlin, B., & Kay, P. (1969). Basic color terms: Their universality and evolution . Berkeley, CA: University of California Press.

Google Scholar  

Gordon, P. (2004). Numerical cognition without words: Evidence from Amazonia. Science, 306 , 496–499.

PubMed   Google Scholar  

Heider, E., & Olivier, D. (1972). The structure of the color space in naming and memory for two languages. Cognitive Psychology, 3 , 337–354.

Hunt, E., & Agnoli, F. (1991). The Whorfian hypothesis: A cognitive approach. Psychological Review, 98 , 377–389.

Pica, P., Lemer, C., & Izard, V. (2004). Exact and approximate calculation in an Amazonian indigene group with a reduced number lexicon. Science, 306 , 499–503.

Roberson, D., Davies, I., & Davidoff, J. (2000). Colour categories are not universal: Replications and new evidence from a Stone-Age culture. Journal of Experimental Psychology: Genera, 129 , 369–398.

Whorf, B. L. (1956). Language, thought, and reality . Cambridge, MA: MIT Press.

Download references

Author information

Authors and affiliations.

Department of Psychology, Oklahoma State University, 116 North Murray Hall, Stillwater, OK, 74078, USA

Seongwon Yun & Shelia M. Kennison

You can also search for this author in PubMed   Google Scholar

Editor information

Editors and affiliations.

Neurology, Learning and Behavior Center, 230 South 500 East, Suite 100, Salt Lake City, Utah, 84102, USA

Sam Goldstein Ph.D.

Department of Psychology MS 2C6, George Mason University, Fairfax, VA, 22030, USA

Jack A. Naglieri Ph.D. ( Professor of Psychology ) ( Professor of Psychology )

Rights and permissions

Reprints and permissions

Copyright information

© 2011 Springer Science+Business Media, LLC

About this entry

Cite this entry.

Yun, S., Kennison, S.M. (2011). Whorfian Hypothesis. In: Goldstein, S., Naglieri, J.A. (eds) Encyclopedia of Child Behavior and Development. Springer, Boston, MA. https://doi.org/10.1007/978-0-387-79061-9_3087

Download citation

DOI : https://doi.org/10.1007/978-0-387-79061-9_3087

Publisher Name : Springer, Boston, MA

Print ISBN : 978-0-387-77579-1

Online ISBN : 978-0-387-79061-9

eBook Packages : Behavioral Science

Share this entry

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Publish with us

Policies and ethics

  • Find a journal
  • Track your research

SEP home page

  • Table of Contents
  • Random Entry
  • Chronological
  • Editorial Information
  • About the SEP
  • Editorial Board
  • How to Cite the SEP
  • Special Characters
  • Advanced Tools
  • Support the SEP
  • PDFs for SEP Friends
  • Make a Donation
  • SEPIA for Libraries
  • Entry Contents

Bibliography

Academic tools.

  • Friends PDF Preview
  • Author and Citation Info
  • Back to Top

Philosophy of Linguistics

Philosophy of linguistics is the philosophy of science as applied to linguistics. This differentiates it sharply from the philosophy of language, traditionally concerned with matters of meaning and reference.

As with the philosophy of other special sciences, there are general topics relating to matters like methodology and explanation (e.g., the status of statistical explanations in psychology and sociology, or the physics-chemistry relation in philosophy of chemistry), and more specific philosophical issues that come up in the special science at issue (simultaneity for philosophy of physics; individuation of species and ecosystems for the philosophy of biology). General topics of the first type in the philosophy of linguistics include:

  • What the subject matter is,
  • What the theoretical goals are,
  • What form theories should take, and
  • What counts as data.

Specific topics include issues in language learnability, language change, the competence-performance distinction, and the expressive power of linguistic theories.

There are also topics that fall on the borderline between philosophy of language and philosophy of linguistics: of “linguistic relativity” (see the supplement on the linguistic relativity hypothesis in the Summer 2015 archived version of the entry on relativism ), language vs. idiolect , speech acts (including the distinction between locutionary, illocutionary, and perlocutionary acts), the language of thought, implicature, and the semantics of mental states (see the entries on analysis , semantic compositionality , mental representation , pragmatics , and defaults in semantics and pragmatics ). In these cases it is often the kind of answer given and not the inherent nature of the topic itself that determines the classification. Topics that we consider to be more in the philosophy of language than the philosophy of linguistics include intensional contexts, direct reference, and empty names (see the entries on propositional attitude reports , intensional logic , rigid designators , reference , and descriptions ).

This entry does not aim to provide a general introduction to linguistics for philosophers; readers seeking that should consult a suitable textbook such as Akmajian et al. (2010) or Napoli (1996). For a general history of Western linguistic thought, including recent theoretical linguistics, see Seuren (1998). Newmeyer (1986) is useful additional reading for post-1950 American linguistics. Tomalin (2006) traces the philosophical, scientific, and linguistic antecedents of Chomsky’s magnum opus (1955/1956; published 1975), and Scholz and Pullum (2007) provide a critical review. Articles that have focused on the philosophical implications of generative linguistics include Ludlow (2011) and Rey (2020). For recent articles on the philosophy of linguistics more generally, Itkonen (2013) discusses various aspects of the field from its early Greek beginnings, Pullum (2019) details debates that have engaged philosophers from 1945 to 2015, and Nefdt (2019a) discusses connections with contemporary issues in the philosophy of science.

1.1 The Externalists

1.2 the emergentists, 1.3 the essentialists, 1.4 comparing the three approaches, 2.1 competence and performance, 2.2 ‘i-language’ and ‘e-language’, 2.3 the faculty of language in narrow and broad senses, 2.4 linguistic ontology, 2.5 components of linguistic theories, 3.1 acrimony over linguistic intuitions, 3.2 grammaticality and acceptability judgments, 3.3 assessing degrees of acceptability, 3.4 informal and experimental elicitation, 3.5 what informal methods actually are, 3.6 corpus data, 4.1 linguistic nativism, 4.2 language learnability, 5.1 phylogenetic emergence, 5.2 historical evolution, other internet resources, related entries.

  • Supplement on Whorfianism

1. Three Approaches to Linguistic Theorizing: Externalism, Emergentism, and Essentialism

The issues we discuss have been debated with vigor and sometimes venom. Some of the people involved have had famous exchanges in the linguistics journals, in the popular press, and in public forums. To understand the sharp disagreements between advocates of the approaches it may be useful to have a sketch of the dramatis personae before us, even if it is undeniably an oversimplification.

We see three tendencies or foci, divided by what they take to be the subject matter, the approach they advocate for studying it, and what they count as an explanation. We characterize them roughly in Table 1.

Table 1. Three Approaches to the Study of Language

A broad and varied range of distinct research projects can be pursued within any of these approaches; one advocate may be more motivated by some parts of the overall project than others are. So the tendencies should not be taken as sharply honed, well-developed research programs or theories. Rather, they provide background biases for the development of specific research programs—biases which sometimes develop into ideological stances or polemical programs or lead to the branching off of new specialisms with separate journals. In the judgment of Phillips (2010), “Dialog between adherents of different approaches is alarmingly rare.”

The names we have given these approaches are just mnemonic tags, not descriptions. The Externalists, for example, might well have been called ‘structural descriptivists’ instead, since they tend to be especially concerned to develop models that can be used to predict the structure of natural language expressions. The Externalists have long been referred to by Essentialists as ‘empiricists’ (and sometimes Externalists apply that term to themselves), though this is misleading (see Scholz and Pullum 2006: 60–63): the ‘empiricist’ tag comes with an accusation of denying the role of learning biases in language acquisition (see Matthews 1984, Laurence and Margolis 2001), but that is no part of the Externalists’ creed (see e.g. Elman 1993, Lappin and Shieber 2007).

Emergentists are also sometimes referred to by Essentialists as ‘empiricists’, but they either use the Emergentist label for themselves (Bates et al. 1998, O’Grady 2008, MacWhinney 2005) or call themselves ‘usage-based’ linguists (Barlow and Kemmer 2002, Tomasello 2003) or ‘construction grammarians’ (Goldberg 1995, Croft 2001). Newmeyer (1991), like Tomasello, refers to the Essentialists as ‘formalists’, because of their tendency to employ abstractions, and to use tools from mathematics and logic.

Despite these terminological inconsistencies, we can look at what typical members of each approach would say about their vision of linguistic science, and what they say about the alternatives. Many of the central differences between these approaches depend on what proponents consider to be the main project of linguistic theorizing, and what they count as a satisfying explanation.

Many researchers—perhaps most—mix elements from each of the three approaches. For example, if Emergentists are to explain the syntactic structure of expressions by appeal to facts about the nature of the use of symbols in human communication, then they will presuppose a great deal of Externalist work in describing linguistic patterns, and those Externalists who work on computational parsing systems frequently use (at least as a starting point) rule systems and ‘structural’ patterns worked out by Essentialists. Certainly, there are no logical impediments for a researcher with one tendency from simultaneously pursuing another; these approaches are only general centers of emphasis.

If one assumes, with the Externalists, that the main goal of a linguistic theory is to develop accurate models of the structural properties of the speech sounds, words, phrases, and other linguistic items, then the clearly privileged information will include corpora (written and oral)—bodies of attested and recorded language use (suitably idealized). The goal is to describe how this public record exhibits certain (perhaps non-phenomenal) patterns that are projectable.

American structural linguistics of the 1920s to 1950s championed the development of techniques for using corpora as a basis for developing structural descriptions of natural languages, although such work was really not practically possible until the wide-spread availability of cheap, powerful, and fast computers. André Martinet (1960: 1) notes that one of the basic assumptions of structuralist approaches to linguistics is that “nothing may be called ‘linguistic’ that is not manifest or manifested one way or another between the mouth of the speaker and the ears of the listener”. He is, however, quick to point out that “this assumption does not entail that linguists should restrict their field of research to the audible part of the communication process—speech can only be interpreted as such, and not as so much noise, because it stands for something else that is not speech.”

American structuralists—Leonard Bloomfield in particular—were attacked, sometimes legitimately and sometimes illegitimately, by certain factions in the Essentialist tradition. For example, it was perhaps justifiable to criticize Bloomfield for adopting a nominalist ontology as popularized by the logical empiricists. But he was later attacked by Essentialists for holding anti-mentalist views about linguistics, when it is arguable that his actual view was that the science of linguistics should not commit itself to any particular psychological theory. (He had earlier been an enthusiast for the mentalist and introspectionist psychology of Wilhelm Wundt; see Bloomfield 1914.)

Externalism continues to thrive within computational linguistics, where the American structuralist vison of studying language through automatic analysis of corpora has enjoyed a recrudescence, and very large, computationally searchable corpora are being used to test hypotheses about the structure of languages (see Sampson 2001, chapter 1, for discussion).

Emergentists aim to explain the capacity for language in terms of non-linguistic human capacities: thinking, communicating, and interacting. Edward Sapir expressed a characteristic Emergentist theme when he wrote:

Language is primarily a cultural or social product and must be understood as such… It is peculiarly important that linguists, who are often accused, and accused justly, of failure to look beyond the pretty patterns of their subject matter, should become aware of what their science may mean for the interpretation of human conduct in general. (Sapir 1929: 214)

The “pretty patterns” derided here are characteristic of structuralist analyses. Sociolinguistics, which is much closer in spirit to Sapir’s project, studies the influence of social and linguistic structure on each other. One particularly influential study, Labov (1966), examines the influence of social class on language variation. Other sociolinguists examine the relation between status within a group on linguistic innovation (Eckert 1989). This interest in variation within languages is characteristic of Emergentist approaches to the study of language.

Another kind of Emergentist, like Tomasello (2003), will stress the role of theory of mind and the capacity to use symbols to change conspecifics’ mental states as uniquely human preadaptations for language acquisition, use, and invention. MacWhinney (2005) aims to explain linguistic phenomena (such as phrase structure and constraints on long distance dependencies) in terms of the way conversation facilitates accurate information-tracking and perspective-switching.

Functionalist research programs generally fall within the broad tendency to approach the study of language as an Emergentist. According to one proponent:

The functionalist view of language [is] as a system of communicative social interaction… Syntax is not radically arbitrary, in this view, but rather is relatively motivated by semantic, pragmatic, and cognitive concerns. (Van Valin 1991, quoted in Newmeyer 1991: 4; emphasis in original)

And according to Russ Tomlin, a linguist who takes a functionalist approach:

Syntax is not autonomous from semantics or pragmatics…the rejection of autonomy derives from the observation that the use of particular grammatical forms is strongly linked, even deterministically linked, to the presence of particular semantic or pragmatic functions in discourse. (Tomlin 1990, quoted by Newmeyer (1991): 4)

The idea that linguistic form is autonomous, and more specifically that syntactic form (rather than, say, phonological form) is autonomous, is a characteristic theme of the Essentialists. And the claims of Van Valin and Tomlin to the effect that syntax is not independent of semantics and pragmatics might tempt some to think that Emergentism and Essentialism are logically incompatible. But this would be a mistake, since there are a large number of nonequivalent autonomy of form theses.

Even in the context of trying to explain what the autonomy thesis is, Newmeyer (1991: 3) talks about five formulations of the thesis, each of which can be found in some Essentialists’ writings, without (apparently) realizing that they are non-equivalent. One is the relatively strong claim that the central properties of linguistic form must not be defined with essential reference to “concepts outside the system”, which suggests that no primitives in linguistics could be defined in psychological or biological terms. Another takes autonomy of form to be a normative claim: that linguistic concepts ought not to be defined or characterized in terms of non-linguistic concepts. The third and fourth versions are ontological: one denies that central linguistic concepts should be ontologically reduced to non-linguistic ones, and the other denies that they can be. And in the fifth version the autonomy of syntax is taken to deny that syntactic patterning can be explained in terms of meaning or discourse functions.

For each of these versions of autonomy, there are Essentialists who agree with it. Probably the paradigmatic Essentialist agrees with them all. But Emergentists need not disagree with them all. Paradigmatic functionalists like Tomlin, Van Valin and MacWhinney could in principle hold that the explanation of syntactic form, for example, will ultimately be in terms of discourse functions and semantics, but still accept that syntactic categories cannot be reduced to non-linguistic ones.

If Leonard Bloomfield is the intellectual ancestor of Externalism, and Sapir the father of Emergentism, then Noam Chomsky is the intellectual ancestor of Essentialism. The researcher with predominantly Essentialist inclinations aims to identify the intrinsic properties of language that make it what it is. For a huge majority of practitioners of this approach—researchers in the tradition of generative grammar associated with Chomsky—this means postulating universals of human linguistic structure, unlearned but tacitly known, that permit and assist children to acquire human languages. This generative Essentialism has a preference for finding surprising characteristics of languages that cannot be inferred from the data of usage, and are not predictable from human cognition or the requirements of communication.

Rather than being impressed with language variation, as are Emergentists and many Externalists, the generative Essentialists are extremely impressed with the idea that very young children of almost any intelligence level, and just about any social upbringing, acquire language to the same high degree of mastery. From this it is inferred that there must be unlearned features shared by all languages that somehow assist in language acquisition.

A large number of contemporary Essentialists who follow Chomsky’s teaching on this matter claim that semantics and pragmatics are not a central part of the study of language. In Chomsky’s view, “it is possible that natural language has only syntax and pragmatics” (Chomsky 1995: 26); that is, only “internalist computations and performance systems that access them”; semantic theories are merely “part of an interface level” or “a form of syntax” (Chomsky 1992: 223).

Thus, while Bloomfield understood it to be a sensible practical decision to assign semantics to some field other than linguistics because of the underdeveloped state of semantic research, Chomsky appears to think that semantics as standardly understood is not part of the essence of the language faculty at all. (In broad outline, this exclusion of semantics from linguistics comports with Sapir’s view that form is linguistic but content is cultural.)

Although Chomsky is an Essentialist in his approach to the study of language, excluding semantics as a central part of linguistic theory clearly does not follow from linguistic Essentialism (Katz 1980 provides a detailed discussion of Chomsky’s views on semantics). Today there are many Essentialists who do hold that semantics is a component of a full linguistic theory.

For example, many linguists today are interested in the syntax-semantics interface—the relationship between the surface syntactic structure of sentences and their semantic interpretation. This area of interest is generally quite alien to philosophers who are primarily concerned with semantics only, and it falls outside of Chomsky’s syntactocentric purview as well. Linguists who work in the kind of semantics initiated by Montague (1974) certainly focus on the essential features of language (most of their findings appear to be of universal import rather than limited to the semantic rules of specific languages). Useful works to consult to get a sense of the modern style of investigation of the syntax-semantics interface would include Partee (1975), Jacobson (1996), Szabolcsi (1997), Chierchia (1998), Steedman (2000).

The discussion so far has been at a rather high level of abstraction. It may be useful to contrast the three tendencies by looking at how they each would analyze a particular linguistic phenomenon. We have selected the syntax of double-object clauses like Hand the guard your pass (also called ditransitive clauses), in which the verb is immediately followed by a sequence of two noun phrases, the first typically denoting a recipient and the second something transferred. For many such clauses there is an alternative way of expressing roughly the same thing: for Hand the guard your pass there is the alternative Hand your pass to the guard , in which the verb is followed by a single object noun phrase and the recipient is expressed after that by a preposition phrase with to . We will call these recipient-PP clauses.

1.4.1 A typical Essentialist analysis

Larson (1988) offers a generative Essentialist approach to the syntax of double-object clauses. In order to provide even a rough outline of his proposals, it will be very useful to be able to use tree diagrams of syntactic structure. A tree is a mathematical object consisting of a set of points called nodes between which certain relations hold. The nodes correspond to syntactic units; left-right order on the page corresponds to temporal order of utterance between them; and upward connecting lines represent the relation ‘is an immediate subpart of’. Nodes are labeled to show categories of phrases and words, such as noun phrase (NP); preposition phrase (PP); and verb phrase (VP). When the internal structure of some subpart of a tree is basically unimportant to the topic under discussion, it is customary to mask that part with an empty triangle. Consider a simple example: an active transitive clause like (Ai) and its passive equivalent (Aii).

A tree structure for (Ai) is shown in (T1).

A tree structure for sentence (Ai)

In analyses of the sort Larson exemplifies, the structure of an expression is given by a derivation , which consists of a sequence of successively modified trees. Larson calls the earliest ones underlying structures. The last (and least abstract) in the derivation is the surface structure, which captures properties relevant to the way the expression is written and pronounced. The underlying structures are posited in order to better identify syntactic generalizations. They are related to surface structures by a series of operations called transformations (which generative Essentialists typically regard as mentally real operations of the human language faculty).

One of the fundamental operations that a transformation can effect is movement , which involves shifting a part of the syntactic structure of a tree to another location within it. For example, it is often claimed that passive clauses have very much the same kinds of underlying structures as the synonymous active clauses, and thus a passive clause like (Aii) would have an underlying structure much like (T1). A movement transformation would shift the guard toward the end of the clause (and add by ), and another would shift my pass into the position before the verb. In other words, passive clauses look much more like their active counterparts in underlying structure.

In a similar way, Larson proposes that a double-object clause like (B.ii) has the same underlying structure as (B.i).

Moreover, he proposes that the transformational operation of deriving the surface structure of (B.ii) from the underlying structure of (B.i) is essentially the same as the one that derives the surface structure of (A.ii) from the underlying structure of (A.i).

Larson adopts many assumptions from Chomsky (1981) and subsequent work. One is that all NPs have to be assigned Case in the course of a derivation. (Case is an abstract syntactic property, only indirectly related to the morphological case forms displayed by nominative, accusative, and genitive pronouns. Objective Case is assumed to be assigned to any NP in direct object position, e.g., my pass in (T1), and Nominative Case is assigned to an NP in the subject position of a tensed clause, e.g., the guard in (T1).)

He also makes two specific assumptions about the derivation of passive clauses. First, Case assignment to the position immediately after the verb is “suppressed”, which entails that the NP there will not get Case unless it moves to some other position. (The subject position is the obvious one, because there it will receive Nominative Case.) Second, there is an unusual assignment of semantic role to NPs: instead of the subject NP being identified as the agent of the action the clause describes, that role is assigned to an adjunct at the end of the VP (the by -phrase in (A.ii); an adjunct is a constituent with an optional modifying role in its clause rather than a grammatically obligatory one like subject or object).

Larson proposes that both of these points about passive clauses have analogs in the structure of double-object VPs. First, Case assignment to the position immediately after the verb is suppressed; and since Larson takes the preposition to to be the marker of Case, this means in effect that to disappears. This entails that the NP after to will not get Case unless it moves to some other position. Second, there is an unusual assignment of semantic role to NPs: instead of the direct object NP being identified as the entity affected by the action the clause describes, that role is assigned to an adjunct at the end of the VP.

Larson makes some innovative assumptions about VPs. First, he proposes that in the underlying structure of a double-object clause the direct object precedes the verb , the tree diagram being (T2).

A tree diagram for the sentence (B.ii)

This does not match the surface order of words ( showed my pass to the guard ), but it is not intended to: it is an underlying structure. A transformation will move the verb to the left of my pass to produce the surface order seen in (B.i).

Second, he assumes that there are two nodes labeled VP in a double-object clause, and two more labeled V′, though there is only one word of the verb (V) category. (Only the smaller VP and V′ are shown in the partial structure (T2).)

What is important here is that (T2) is the basis for the double-object surface structure as well. To produce that, the preposition to is erased and an additional NP position (for my pass ) is attached to the V′, thus:

Transforming the tree T2 by erasing a preposition and adding a new NP

The additional NP is assigned the affected-entity semantic role. The other NP ( the guard ) does not yet have Case; but Larson assumes that it moves into the NP position before the verb. The result is shown in (T4), where ‘ e ’ marks the empty string left where some words have been moved away:

Positioning the new NP before the verb

Larson assumes that in this position the guard can receive Case. What remains is for the verb to move into a higher V position further to its left, to obtain the surface order:

Moving the verb to a higher V position

The complete sequence of transformations is taken to give a deep theoretical explanation of many properties of (B.i) and (B.ii), including such things as what could be substituted for the two NPs, and the fact there is at least rough truth-conditional equivalence between the two clauses.

The reader with no previous experience of generative linguistics will have many questions about the foregoing sketch (e.g., whether it is really necessary to have the guard after showed in (T3), then the opposite order in (T4), and finally the same order again in (T5)). We cannot hope to answer such questions here; Larson’s paper is extremely rich in further assumptions, links to the previous literature, and additional classes of data that he aims to explain. But the foregoing should suffice to convey some of the flavor of the analysis.

The key point to note is that Essentialists seek underlying symmetries and parallels whose operation is not manifest in the data of language use. For Essentialists, there is positive explanatory virtue in hypothesizing abstract structures that are very far from being inferrable from performance; and the posited operations on those structures are justified in terms of elegance and formal parallelism with other analyses, not through observation of language use in communicative situations.

1.4.2 A typical Emergentist analysis

Many Emergentists are favorably disposed toward the kind of construction grammar expounded in Goldberg (1995). We will use her work as an exemplar of the Emergentist approach. The first thing to note is that Goldberg does not take double-object clauses like (B.ii) to be derived alternants of recipient-PP structures like (B.i), the way Larson does. So she is not looking for a regular syntactic operation that can relate their derivations; indeed, she does not posit derivations at all. She is interested in explaining correlations between syntactic, semantic, and pragmatic aspects of clauses; for example, she asks this question:

How are the semantics of independent constructions related such that the classes of verbs associated with one overlap with the classes of verbs associated with another? (Goldberg 1995: 89)

Thus she aims to explain why some verbs occur in both the double-object and recipient-PP kinds of expression and some do not.

The fundamental notion in Goldberg’s linguistic theory is that of a construction . A construction can be defined very roughly as a way of structurally composing words or phrases—a sort of template—for expressing a certain class of meanings. Like Emergentists in general, Goldberg regards linguistic theory as continuous with a certain part of general cognitive psychological theory; linguistics emerges from this more general theory, and linguistic matters are rarely fully separate from cognitive matters. So a construction for Goldberg has a mental reality: it corresponds to a generalized concept or scenario expressible in a language, annotated with a guide to the linguistic structure of the expression.

Many words will be trivial examples of constructions: a single concept paired with a way of pronouncing and some details about grammatical restrictions (category, inflectional class, etc.); but constructions can be much more abstract and internally complex. The double-object construction, which Goldberg calls the Ditransitive Construction, is a moderately abstract and complex one; she diagrams it thus (p. 50):

Goldberg's Ditransitive Construction

This expresses a set of constraints on how to use English to communicate the idea of a particular kind of scenario. The scenario involves a ternary relation CAUSE-RECEIVE holding between an agent ( agt ), a recipient ( rec ), and a patient ( pat ). PRED is a variable that is filled by the meaning of a particular verb when it is employed in this construction.

The solid vertical lines downward from agt and pat indicate that for any verb integrated into this construction it is required that its subject NP should express the agent participant, and the direct object (OBJ 2 ) should express the patient participant. The dashed vertical line downward from rec signals that the first object (OBJ) may express the recipient but it does not have to—the necessity of there being a recipient is a property of the construction itself, and not every verb demands that it be made explicit who the recipient is. But if there are two objects, the first is obligatorily associated with the recipient role: We sent the builder a carpenter can only express a claim about the sending of a carpenter over to the builder, never the sending of the builder over to where a carpenter is.

When a particular verb is used in this construction, it may have obligatory accompanying NPs denoting what Goldberg calls “profiled participants” so that the match between the participant roles ( agt , rec , pat ) is one-to-one, as with the verb hand . When this verb is used, the agent (‘hander’), recipient (‘handee’), and item transferred (‘handed’) must all be made explicit. Goldberg gives the following diagram of the “composite structure” that results when hand is used in the construction:

Instance of Goldberg's Ditransitive Construction

Because of this requirement of explicit presence, Hand him your pass is grammatical, but * Hand him is not, and neither is * Hand your pass . The verb send , on the other hand, illustrates the optional syntactic expression of the recipient role: we can say Send a text message , which is understood to involve some recipient but does not make the recipient explicit.

The R notation relates to the fact that particular verbs may express either an instance of causing someone to receive something, as with hand , or a means of causing someone to receive something, as with kick : what Joe kicked Bill the ball means is that Joe caused Bill to receive the ball by means of a kicking action.

Goldberg’s discussion covers many subtle ways in which the scenario communicated affects whether the use of a construction is grammatical and appropriate. For example, there is something odd about ? Joe kicked Bill the ball he was trying to kick to Sam : the Ditransitive Construction seems best suited to cases of volitional transfer (rather than transfer as an unexpected side effect of a blunder). However, an exception is provided by a class of cases in which the transfer is not of a physical object but is only metaphorical: That guy gives me the creeps does not imply any volitional transfer of a physical object.

Metaphorical cases are distinguished from physical transfers in other ways as well. Goldberg notes sentences like The music lent the event a festive air , where the music is subject of the verb lend despite the fact that music cannot literally lend anything to anyone.

Goldberg discusses many topics such as metaphorical extension, shading, metonymy, cutting, role merging, and also presents various general principles linking meanings and constructions. One of these principles, the No Synonymy Principle, says that no two syntactically distinct constructions can be both semantically and pragmatically synonymous. It might seem that if any two sentences are synonymous, pairs like this are:

Yet the two constructions cannot be fully synonymous, both semantically and pragmatically, if the No Synonymy Principle is correct. And to support the principle, Goldberg notes purported contrasts such as this:

There is a causation-as-transfer metaphor here, and it seems to be compatible with the double object construction but not with the recipient-PP. So (in Goldberg’s view) the two are not fully synonymous.

It is no part of our aim here to provide a full account of the content of Goldberg’s discussion of double-object clauses. But what we want to highlight is that the focus is not on finding abstract elements or operations of a purely syntactic nature that are candidates for being essential properties of language per se. The focus for Emergentists is nearly always on the ways in which meaning is conveyed, the scenarios that particular constructions are used to communicate, and the aspects of language that connect up with psychological topics like cognition, perception, and conceptualization.

1.4.3 A typical Externalist analysis

One kind of work that is representative of the Externalist tendency is nicely illustrated by Bresnan et al. (2007) and Bresnan and Ford (2010). Bresnan and her colleagues defend the use of corpora—bodies of attested written and spoken texts. One of their findings is that a number of types of expressions that linguists have often taken to be ungrammatical do in fact turn up in actual use. Essentialists and Emergentists alike have often, purely on the basis of intuition, asserted that sentences like John gave Mary a kiss are grammatical but sentences like John gave a kiss to Mary are no, as we see above with Goldberg’s (D)(ii). Bresnan and her colleagues find numerous occurrences of the latter sort on the World Wide Web, and conclude that they are not ungrammatical or even unacceptable, but merely dispreferred.

Bresnan and colleagues used a three-million-word collection of recorded and transcribed spontaneous telephone conversations known as the Switchboard corpus to study the double-object and recipient-PP constructions. They first annotated the utterances with indications of a number of factors that they thought might influence the choice between the double-object and recipient-PP constructions:

  • Discourse accessibility of NPs: does a particular NP refer to something already mentioned, or to something new to the discourse?
  • Relative lengths of NPs: what is the difference in number of words between the recipient NP and the transferred-item NP?
  • Definiteness: are the recipient and transferred-item NPs definite like the bishop or indefinite like some members
  • Animacy: do the recipient and transferred-item NPs denote animate beings or inanimate things?
  • Pronominality: are the recipient and transferred-item NPs pronouns?
  • Number: are the recipient and transferred-item NPs singular or plural?
  • person: are the recipient and transferred-item NPs first-person or second-person pronouns, or third person?

They also coded the verb meanings by assigning them to half a dozen semantic categories:

  • Abstract senses ( give it some thought );
  • Transfer of possession ( give him an armband );
  • Future transfer of possession ( I owe you a dollar );
  • Prevention of possession ( They denied me my rights );
  • Communication verb sense ( tell me your name ).

They then constructed a statistical model of the corpus: a mathematical formula expressing, for each combination of the factors listed above, the ratio of the probabilities of the double object and the recipient-PP. (To be precise, they used the natural logarithm of the ratio of p to 1 − p , where p is the probability of a double-object or recipient-PP in the corpus being of the double-object form.) They then used logistic regression to predict the probability of fit to the data.

To determine how well the model generalized to unseen data, they divided the data randomly 100 times into a training set and a testing set, fit the model parameters on each training set, and scored its predictions on the unseen testing set. The average percent of correct predictions on unseen data was 92%. All components of the model except number of the recipient NP made a statistically significant difference—almost all at the 0.001 level.

What this means is that knowing only the presence or absence of the sort of factors listed above they were reliably able to predict whether double-object or recipient-PP structures would be used in a given context, with a 92% score accuracy rate.

The implication is that the two kinds of structure are not interchangeable: they are reliably differentiated by the presence of other factors in the texts in which they occur.

They then took the model they had generated for the telephone speech data and applied it to a corpus of written material: the Wall Street Journal corpus (WSJ), a collection of 1987–9 newspaper copy, only roughly edited. The main relevant difference with written language is that the language producer has more opportunity to reflect thoughtfully on how they are going to phrase things. It was reasonable to think that a model based on speech data might not transfer well. But instead the model had 93.5% accuracy. The authors conclude is that “the model for spoken English transfers beautifully to written”. The main difference between the corpora was found to be a slightly higher probability of the recipient-PP structure in written English.

In a very thorough subsequent study, Bresnan and Ford (2010) show that the results also correlate with native speakers’ metalinguistic judgments of naturalness for sentence structures, and with lexical decision latencies (speed of deciding whether the words in a text were genuine English words or not), and with a sentence completion task (choosing the most natural of a list of possible completions of a partial sentence). The results of these experiments confirmed that their model predicted participants’ performance.

Among the things to note about this work is that it was all done on directly recorded performance data: transcripts of people speaking to each other spontaneously on the phone in the case of the Switchboard corpus, stories as written by newspaper journalists in the case of WSJ, measured responses of volunteer subjects in a laboratory in the case of the psycholinguistic experiments of Bresnan and Ford (2010). The focus is on identifying the factors in linguistic performance that permit accurate prediction of future performance, and the methods of investigation have a replicability and checkability that is familiar in the natural sciences.

However, we should make it clear that the work is not some kind of close-to-the-ground collecting and classifying of instances. The models that Bresnan and her colleagues develop are sophisticated mathematical abstractions, very far removed from the records of utterance tokens. They claim that these models “allow linguistic theory to solve more difficult problems than it has in the past, and to build convergent projects with psychology, computer science, and allied fields of cognitive science” (Bresnan et al. 2007: 69).

1.4.4 Conclusion

It is important to see that the contrast we have drawn here is not just between three pieces of work that chose to look at different aspects of the phenomena associated with double-object sentences. It is true that Larson focuses more on details of tree structure, Goldberg more on subtle differences in meaning, and Bresnan et al. on frequencies of occurrence. But that is not what we are pointing to. What we want to stress is that we are illustrating three different broad approaches to language that regard different facts as likely to be relevant, and make different assumptions about what needs to be accounted for, and what might count as an explanation.

Larson looks at contrasts between different kinds of clause with different meanings and see evidence of abstract operations affecting subtle details of tree structure, and parallelism between derivational operations formerly thought distinct.

Goldberg looks at the same facts and sees evidence not for anything to do with derivations but for the reality of specific constructions—roughly, packets of syntactic, semantic, and pragmatic information tied together by constraints.

Bresnan and her colleagues see evidence that readily observable facts about speaker behavior and frequency of word sequences correlate closely with certain lexical, syntactic, and semantic properties of words.

Nothing precludes defenders of any of the three approaches from paying attention to any of the phenomena that the other approaches attend to. There is ample opportunity for linguists to mix aspects of the three approaches in particular projects. But in broad outline there are three different tendencies exhibited here, with stereotypical views and assumptions roughly as we laid them out in Table 1.

2. The Subject Matter of Linguistic Theories

The complex and multi-faceted character of linguistic phenomena means that the discipline of linguistics has a whole complex of distinguishable subject matters associated with different research questions. Among the possible topics for investigation are these:

  • the capacity of humans to acquire, use, and invent languages;
  • the abstract structural patterns (phonetic, morphological, syntactic, or semantic) found in a particular language under some idealization;
  • systematic structural manifestations of the use of some particular language;
  • the changes in a language or among languages across time;
  • the psychological functioning of individuals who have successfully acquired particular languages;
  • the psychological processes underlying speech or linguistically mediated thinking in humans;
  • the evolutionary origin of (i), and/or (ii).

There is no reason for all of the discipline of linguistics to converge on a single subject matter, or to think that the entire field of linguistics cannot have a diverse range of subject matters. To give a few examples:

  • The influential Swiss linguist Ferdinand de Saussure (1916) distinguished between langue , a socially shared set of abstract conventions (compare with (ii)) and parole , the particular choices made by a speaker deploying a language (compare (iii)).
  • The anthropological linguist Edward Sapir (1921, 1929) thought that human beings have a seemingly species-universal capacity to acquire and use languages (compare (i)), but his own interest was limited to the systematic structural features of particular languages (compare (ii)) and the psychological reality of linguistic units such as the phoneme (an aspect of (vi)), and the psychological effects of language and thought (an aspect of (v)).
  • Bloomfield (1933) showed a strong interest in historical linguistic change (compare (iv)), distinguishing that sharply (much as Saussure did) from synchronic description of language structure ((ii) again) and language use (compare (iii)), arguing that the study of (iv) presupposed (vi).
  • Bloomfield famously eschewed all dualistic mentalistic approaches to the study of language, but since he rejected them on materialist ontological grounds, his rejection of mentalism was not clearly a rejection of (vi) or (vii): his attempt to cast linguistics in terms of stimulus-response psychology indicates that he was sympathetic to the Weissian psychology of his time and accepted that linguistics might have psychological subject matter.
  • Zellig Harris, on the other hand, showed little interest in the psychology of language, concentrating on mathematical techniques for tackling (ii).

Most saliently of all, Harris’s student Chomsky reacted strongly against indifference toward the mind, and insisted that the principal subject matter of linguistics was, and had to be, a narrow psychological version of (i), and an individual, non-social, and internalized conception of (ii).

In the course of advancing his view, Chomsky introduced a number of novel pairs of terms into the linguistics literature: competence vs. performance (Chomsky 1965); ‘I-language’ vs. ‘E-language’ (Chomsky 1986); the faculty of language in the narrow sense vs. the and faculty of language in the broad sense (the ‘FLN’ and ‘FLB’ of Hauser et al. 2002). Because Chomsky’s terminological innovations have been adopted so widely in linguistics, the focus of sections 2.1–2.3 will be to examine the use of these expressions as they were introduced into the linguistics literature and consider their relation to (i)–(vii).

Essentialists invariably distinguish between what Chomsky (1965) called competence and performance . Competence is what knowing a language confers: a tacit grasp of the structural properties of all the sentences of a language. Performance involves actual real-time use, and may diverge radically from the underlying competence, for at least two reasons: (a) an attempt to produce an utterance may be perturbed by non-linguistic factors like being distracted or interrupted, changing plans or losing attention, being drunk or having a brain injury; or (b) certain capacity limits of the mechanisms of perception or production may be overstepped.

Emergentists tend to feel that the competence/performance distinction sidelines language use too much. Bybee and McClelland put it this way:

One common view is that language has an essential and unique inner structure that conforms to a universal ideal, and what people say is a potentially imperfect reflection of this inner essence, muddied by performance factors. According to an opposing view…language use has a major impact on language structure. The experience that users have with language shapes cognitive representations, which are built up through the application of general principles of human cognition to linguistic input. The structure that appears to underlie language use reflects the operation of these principles as they shape how individual speakers and hearers represent form and meaning and adapt these forms and meanings as they speak. (Bybee and McClelland 2005: 382)

And Externalists are often concerned to describe and explain not only language structure, but also the workings of processing mechanisms and the etiology of performance errors.

However, every linguist accepts that some idealization away from the speech phenomena is necessary. Emergentists and Externalists are almost always happy to idealize away from sporadic speech errors. What they are not so keen to do is to idealize away from limitations on linguistic processing and the short-term memory on which it relies. Acceptance of a thoroughgoing competence/performance distinction thus tends to be a hallmark of Essentialist approaches, which take the nature of language to be entirely independent of other human cognitive processes (though of course capable of connecting to them).

The Essentialists’ practice of idealizing away from even psycholinguistically relevant factors like limits on memory and processing plays a significant role in various important debates within linguistics. Perhaps the most salient and famous is the issue of whether English is a finite-state language.

The claim that English is not accepted by any finite-state automaton can only be supported by showing that every grammar for English has center- embedding to an unbounded depth (see Levelt 2008: 20–23 for an exposition and proof of the relevant theorem, originally from Chomsky 1959). But even depth-3 center-embedding of clauses (a clause interrupting a clause that itself interrupts a clause) is in practice extraordinarily hard to process. Hardly anyone can readily understand even semantically plausible sentences like Vehicles that engineers who car companies trust build crash every day . And such sentences virtually never occur, even in writing. Karlsson (2007) undertakes an extensive examination of available textual material, and concludes that depth-3 center-embeddings are vanishingly rare, and no genuine depth-4 center-embedding has ever occurred at all in naturally composed text. He proposes that there is no reason to regard center-embedding as grammatical beyond depth 3 (and for spoken language, depth 2). Karlsson is proposing a grammar that stays close to what performance data can confirm; the standard Essentialist view is that we should project massively from what is observed, and say that depth- n center-embedding is fully grammatical for all n .

Chomsky (1986) introduced into the linguistics literature two technical notions of a language: ‘E-Language’ and ‘I-Language’. He deprecates the former as either undeserving of study or as a fictional entity, and promotes the latter as the only scientifically respectable object of study for a serious linguistics.

2.2.1 ‘E-language’

Chomsky’s notion ‘E-language’ is supposed to suggest by its initial ‘E’ both ‘extensional’ (concerned with which sentences happen to satisfy a definition of a language rather than with what the definition says) and ‘external’ (external to the mind, that is, non-mental). The dismissal of E-language as an object of study is aimed at critics of Essentialism—many but not all of those critics falling within our categories of Externalists and Emergentists.

Extensional . First, there is an attempt to impugn the extensional notion of a language that is found in two radically different strands of Externalist work. Some Externalist investigations are grounded in the details of attested utterances (as collected in corpora), external to human minds. Others, with mathematical or computational interests, sometimes idealize languages as extensionally definable objects (typically infinite sets of strings) with a certain structure, independently of whatever device might be employed to characterize them. A set of strings of words either is or is not regular (finite-state), either is or is not recursive (decidable), etc., independently of forms of grammar statement. Chomsky (1986) basically dismissed both corpus-based work and mathematical linguistics simply on the grounds that they employ an extensional conception of language that is, a conception that removes the object of study from having an essential connection with the mental.

External . Second, a distinct meaning based on ‘external’ was folded into the neologism ‘E-language’ to suggest criticism of any view that conceives of a natural language as a public, intersubjectively accessible system used by a community of people (often millions of them spread across different countries). Here, the objection is that languages as thus conceived have no clear criteria of individuation in terms of necessary and sufficient conditions. On this conception, the subject matter of interest is a historico-geographical entity that changes as it is transmitted over generations, or over mountain ranges. Famously, for example, there is a gradual valley-to-valley change in the language spoken between southeastern France and northwestern Italy such that each valley’s speakers can understand the next. But the far northwesterners clearly speak French and the far southeasterners clearly speak Italian. It is the politically defined geographical border, not the intrinsic properties of the dialects, that would encourage viewing this continuum as two different languages.

Perhaps the most famous quotation by any linguist is standardly attributed to Max Weinreich (1945): ‘A shprakh iz a dialekt mit an armey un flot’ (‘A language is a dialect with an army and navy’; he actually credits the remark to an unnamed student). The implication is that E-languages are defined in terms of non-linguistic, non-essential properties. Essentialists object that a scientific linguistics cannot tolerate individuating French and Italian in a way that is subject to historical contingencies of wars and treaties (after all, the borders could have coincided with a different hill or valley had some battle had a different outcome).

Considerations of intelligibility fare no better. Mutual intelligibility between languages is not a transitive relation, and sometimes the intelligibility relation is not even symmetric (smaller, more isolated, or less prestigious groups often understand the dialects of larger, more central, or higher-prestige groups when the converse does not hold). So these sociological facts cannot individuate languages either.

Chomsky therefore concludes that languages cannot be defined or individuated extensionally or mind-externally, and hence the only scientifically interesting conception of a ‘language’ is the ‘I-language’ view (see for example Chomsky 1986: 25; 1992; 1995 and elsewhere). Chomsky says of E-languages that “all scientific approaches have simply abandoned these elements of what is called ‘language’ in common usage” (Chomsky 1988, 37); and “we can define E-language in one way or another or not at all, since the concept appears to play no role in the theory of language” (Chomsky 1986: 26; in saying that it appears to play no role in the theory of language, here he means that it plays no role in the theory he favours).

This conclusion may be bewildering to non-linguists as well as non-Essentialists. It is at odds with what a broad range of philosophers have tacitly assumed or explicitly claimed about language or languages: ‘[A language] is a practice in which people engage…it is constituted by rules which it is part of social custom to follow’ (Dummett 1986: 473–473); ‘Language is a set of rules existing at the level of common knowledge’ and these rules are ‘norms which govern intentional social behavior’ (Itkonen 1978: 122), and so on. Generally speaking, those philosophers influenced by Wittgenstein also take the view that a language is a social-historical entity. But the opposite view has become a part of the conceptual underpinning of linguistics for many Essentialists.

Failing to have precise individuation conditions is surely not a sufficient reason to deny that an entity can be studied scientifically. ‘Language’ as a count noun in the extensional and socio-historical sense is vague, but this need not be any greater obstacle to theorizing about them than is the vagueness of other terms for historical entities without clear individuation conditions, like ‘species’ and ‘individual organism’ in biology.

At least some Emergentist linguists, and perhaps some Externalists, would be content to say that languages are collections of social conventions, publicly shared, and some philosophers would agree (see Millikan 2003, for example, and Chomsky 2003 for a reply). Lewis (1969) explicitly defends the view that language can be understood in terms of public communications, functioning to solve coordination problems within a group (although he acknowledges that the coordination could be between different temporal stages of one individual, so language use by an isolated person is also intelligible; see the appendix “Lewis’s Theory of Languages as Conventions” in the entry on idiolects , for further discussion of Lewis). What Chomsky calls E-languages, then, would be perfectly amenable to linguistic or philosophical study. Santana (2016) makes a similar argument in terms of scientific idealization. He argues that since all sciences idealize their targets, Chomsky needs to do more to show why idealizations concerning E-languages are illicit (see also Stainton 2014).

2.2.2 ‘I-language’

Chomsky (1986) introduced the neologism ‘I-language’ in part to disambiguate the word ‘grammar’. In earlier generative Essentialist literature, ‘grammar’ was (deliberately) ambiguous between (i) the linguist’s generative theory and (ii) what a speaker knows when they know a language. ‘I-language’ can be regarded as a replacement for Bever’s term ‘psychogrammar’ (see also George 1989): it denotes a mental or psychological entity (not a grammarian’s description of a language as externally manifested).

I-language is first discussed under the sub-heading of ‘internalized language’ to denote linguistic knowledge. Later discussion in Chomsky 1986 and 1995 makes it clear that the ‘I’ of ‘I-language’ is supposed to suggest at least three English words: ‘individual’, ‘internal’, and ‘intensional’. And Chomsky emphasizes that the neologism also implies a kind of realism about speakers’ knowledge of language.

Individual . A language is claimed to be strictly a property of individual human beings—not groups. The contrast is between the idiolect of a single individual, and a dialect or language of a geographical, social, historical, or political group. I-languages are properties of the minds of individuals who know them.

Internal . As generative Essentialists see it, your I-language is a state of your mind/brain. Meaning is internal—indeed, on Chomsky’s conception, an I-language

is a strictly internalist, individualist approach to language, analogous in this respect to studies of the visual system. If the cognitive system of Jones’s language faculty is in state L, we will say that Jones has the I-language L. (Chomsky 1995: 13)

And he clarifies the sense in which an I-language is internal by appealing to an analogy with the way the study of vision is internal:

The same considerations apply to the study of visual perception along lines pioneered by David Marr, which has been much discussed in this connection. This work is mostly concerned with operations carried out by the retina; loosely put, the mapping of retinal images to the visual cortex. Marr’s famous three levels of analysis—computational, algorithmic, and implementation—have to do with ways of construing such mappings. Again, the theory applies to a brain in a vat exactly as it does to a person seeing an object in motion. (Chomsky 1995: 52)

Thus, while the speaker’s I-language may be involved in performing operations over representations of distal stimuli—representations of other speaker’s utterances—I-languages can and should be studied in isolation from their external environments.

Although Chomsky sometimes refers to this narrow individuation of I-languages as ‘individual’, he clearly claims that I-languages are individuated in isolation from both speech communities and other aspects of the broadly conceived natural environment:

Suppose Jones is a member of some ordinary community, and J is indistinguishable from him except that his total experience derives from some virtual reality design; or let J be Jones’s Twin in a Twin-Earth scenario. They have had indistinguishable experiences and will behave the same way (in so far as behavior is predictable at all); they have the same internal states. Suppose that J replaces Jones in the community, unknown to anyone except the observing scientist. Unaware of any change, everyone will act as before, treating J as Jones; J too will continue as before. The scientist seeking the best theory of all of this will construct a narrow individualist account of Jones, J, and others in the community. The account omits nothing… (Chomsky 1995: 53–54)

This passage can also be seen as suggesting a radically intensionalist conception of language.

Intensional . The way in which I-languages are ‘intensional’ for Chomsky needs a little explication. The concept of intension is familiar in logic and semantics, where ‘intensional’ contrasts with ‘extensional’. The extension of a predicate like blue is simply the set of all blue objects; the intension is the function that picks out in a given world the blue objects contained therein. In a similar way, the extension of a set can be distinguished from an intensional description of the set in terms of a function: the set of integer squares is {1, 4, 9, 16, 25, 36, …}, and the intension could be given in terms of the one-place function f such that f ( n ) = n × n . One difference between the two accounts of squaring is that the intensional one could be applied to a different domain (any domain on which the ‘×’ operation is defined: on the rationals rather than the integers, for example, the extension of the identically defined function is a different and larger set containing infinitely many fractions).

In an analogous way, a language can be identified with the set of all and only its expressions (regardless of what sort of object an expression is: a word sequence, a tree structure, a complete derivation, or whatever), which is the extensional view; but it can also be identified intensionally by means of a recipe or formal specification of some kind—what linguists call a grammar. Ludlow (2011) considers the first I (individual) to be the weakest link and thus the most expendable. He argues in its stead for a concept of a “Ψ-language” which allows for the possibility of the I-language relating to external objects either constitutively or otherwise.

In natural language semantics, an intensional context is one where substitution of co-extensional terms fails to preserve truth value ( Scott is Scott is true, and Scott is the author of Waverley is true, but the truth of George knows that Scott is Scott doesn’t guarantee the truth of George knows that Scott is the author of Waverly , so knows that establishes an intensional context).

Chomsky claims that the truth of an I-language attribution is not preserved by substituting terms that have the same extension. That is, even when two human beings do not differ at all on what expressions are grammatical, it may be false to say that they have the same I-language. Where H is a human being and L is a language (in the informal sense) and R is the relation of knowing (or having, or using) that holds between a human being and a language, Chomsky holds, in effect, that R establishes an intensional context in statements of the theory:

[F]or H to know L is for H to have a certain I-language. The statements of the grammar are statements of the theory of mind about the I-language, hence structures of the brain formulated at a certain level of abstraction from mechanisms. These structures are specific things in the world, with their properties… The I-language L may be the one used by a speaker but not the I-language L′ even if the two generate the same class of expressions (or other formal objects) … L′ may not even be a possible human I-language, one attainable by the language faculty. (Chomsky 1986: 23)

The idea is that two individuals can know (or have, or use) different I-languages that generate exactly the same strings of words, and even give them exactly the same structures. This situation forms the basis of Quine’s (1972) infamous critique of the psychological reality of generative grammars (see Johnson 2015 for a solution in terms of invariance of ‘behaviorally equivalent grammar formalisms, to use Quine’s terminology, see also Nefdt 2021 for a similar resolution in terms of structural realism in the philosophy of science).

The generative Essentialist conception of an I-language is antithetical to Emergentist research programs. If the fundamental explanandum of scientific linguistics is how actual linguistic communication takes place, one must start by looking at both internal (psychological) and external (public) practices and conventions in virtue of which it occurs, and consider the effect of historical and geographic contingencies on the relevant underlying processes. That would not rule out ‘I-language’ as part of the explanans; but some Emergentists seem to be fictionalists about I-languages, in an analogous sense to the way that Chomsky is a fictionalist about E-languages. Emergentists do not see a child as learning a generative grammar, but as learning how to use a symbolic system for propositional communication. On this view grammars are mere artifacts that are developed by linguists to codify aspects of the relevant systems, and positing an I-language amounts to projecting the linguist’s codification illegitimately onto human minds (see, for example, Tomasello 2003).

The I-language concept brushes aside certain phenomena of interest to the Externalists, who hold that the forms of actually attested expressions (sentences, phrases, syllables, and systems of such units) are of interest for linguistics. For example, computational linguistics (work on speech recognition, machine translation, and natural language interfaces to databases) must rely on a conception of language as public and extensional; so must any work on the utterances of young children, or the effects of word frequency on vowel reduction, or misunderstandings caused by road sign wordings. At the very least, it might be said on behalf of this strain of Externalism (along the lines of Soames 1984) that linguistics will need careful work on languages as intersubjectively accessible systems before hypotheses about the I-language that purportedly produces them can be investigated.

It is a highly biased claim that the E-language concept “appears to play no role in the theory of language” (Chomsky 1986: 26). Indeed, the terminological contrast seems to have been invented not to clarify a distinction between concepts but to nudge linguistic research in a particular direction.

In Hauser et al. (2002) (henceforth HCF) a further pair of contrasting terms is introduced. They draw a distinction quite separate from the competence/performance and ‘I-language’/‘E-language’ distinctions: the “language faculty in the narrow sense” (FLN) is distinguished from the “language faculty in the broad sense” (FLB). According to HCF, FLB “excludes other organism-internal systems that are necessary but not sufficient for language (e.g., memory, respiration, digestion, circulation, etc.)” but includes whatever is involved in language, and FLN is some limited part of FLB (p. 1571) This is all fairly vague, but it is clear that FLN and FLB are both internal rather than external, and individual rather than social.

The FLN/FLB distinction apparently aims to address the uniqueness of one component of the human capacity for language rather than (say) the content of human grammars. HCF say (p. 1573) that “Only FLN is uniquely human”; they “hypothesize that most, if not all, of FLB is based on mechanisms shared with nonhuman animals”; and they say:

[T]he computations underlying FLN may be quite limited. In fact, we propose in this hypothesis that FLN comprises only the core computational mechanisms of recursion as they appear in narrow syntax and the mappings to the interfaces. (ibid.)

The components of FLB that HCF hypothesize are not part of FLN are the “sensory-motor” and “conceptual-intentional” systems. The study of the conceptual-intentional system includes investigations of things like the theory of mind; referential vocal signals; whether imitation is goal directed; and the field of pragmatics. The study of the sensory motor system, by contrast, includes “vocal tract length and formant dispersion in birds and primates”; learning of songs by songbirds; analyses of vocal dialects in whales and spontaneous imitation of artificially created sounds in dolphins; “primate vocal production, including the role of mandibular oscillations”; and “[c]ross-modal perception and sign language in humans versus unimodal communication in animals”.

It is presented as an empirical hypothesis that a core property of the FLN is “recursion”:

All approaches agree that a core property of FLN is recursion, attributed to narrow syntax…FLN takes a finite set of elements and yields a potentially infinite array of discrete expressions. This capacity of FLN yields discrete infinity (a property that also characterizes the natural numbers). (HCF, p. 1571)

HCF leave open exactly what the FLN includes in addition to recursion. It is not ruled out that the FLN incorporates substantive universals as well as the formal property of “recursion”. But whatever “recursion” is in this context, it is apparently not domain-specific in the sense of earlier discussions by generative Essentialists, because it is not unique to human natural language or defined over specifically linguistic inputs and outputs: it is the basis for humans’ grasp of the formal and arguably non-natural language of arithmetic (counting, and the successor function), and perhaps also navigation and social relations. It might be more appropriate to say that HCF identify recursion as a cognitive universal, not a linguistic one. And in that case it is difficult to see how the so-called ‘language faculty’ deserves that name: it is more like a faculty for cognition and communication.

This abandonment of linguistic domain-specificity contrasts very sharply with the picture that was such a prominent characteristic of the earlier work on linguistic nativism, popularized in different ways by Fodor (1983), Barkow et al. (1992), and Pinker (1994). And yet the HCF discussion of FLN seems to incline to the view that human language capacities have a unique human (though not uniquely linguistic) essence.

The FLN/FLB distinction provides earlier generative Essentialism with an answer (at least in part) to the question of what the singularity of the human language faculty consists in, and it does so in a way that subsumes many of the empirical discoveries of paleoanthropology, primatology, and ethnography that have been part of highly influential in Emergentist approaches as well as neo-Darwinian Essentialist approaches. A neo-Darwinian Essentialist like Pinker will accept that the language faculty involves recursion, but also will also hold (with Emergentists) that human language capacities originated, via natural selection, for the purpose of linguistic communication.

Thus, over the years, those Essentialists who follow Chomsky closely have changed the term they use for their core subject matter from ‘linguistic competence’ to ‘I-language’ to ‘FLN’, and the concepts expressed by these terms are all slightly different. In particular, what they are counterposed to differs in each case.

The challenge for the generative Essentialist adopting the FLN/FLB distinction as characterized by HCF is to identify empirical data that can support the hypothesis that the FLN “yields discrete infinity”. That will mean answering the question: discrete infinity of what? HCF write that FLN “takes a finite set of elements and yields a potentially infinite array of discrete expressions” (p. 1571), which makes it clear that there must be a recursive procedure in the mathematical sense, perhaps putting atomic elements such as words together to make internally complex elements like sentences (“array” should probably be understood as a misnomer for ‘set’). But then they say, somewhat mystifyingly:

Each of these discrete expressions is then passed to the sensory-motor and conceptual-intentional systems, which process and elaborate this information in the use of language. Each expression is, in this sense, a pairing of sound and meaning. (HCF, p. 1571)

But the sensory-motor and conceptual-intentional systems are concrete parts of the organism: muscles and nerves and articulatory organs and perceptual channels and neuronal activity. How can each one of a “potentially infinite array” be “passed to” such concrete systems without it taking a potentially infinite amount of time? HCF may mean that for any one of the expressions that FLN defines as well-formed (by generating it) there is a possibility of its being used as the basis for a pairing of sound and meaning. This would be closer to the classical generative Essentialist view that the grammar generates an infinite set of structural descriptions; but it is not what HCF say.

At root, HCF is a polemical work intended to identify the view it promotes as valuable and all other approaches to linguistics as otiose.

In the varieties of modern linguistics that concern us here, the term “language” is used quite differently to refer to an internal component of the mind/brain (sometimes called internal language or I-language).… However, this biologically and individually grounded usage still leaves much open to interpretation (and misunderstanding). For example, a neuroscientist might ask: What components of the human nervous system are recruited in the use of language in its broadest sense? Because any aspect of cognition appears to be, at least in principle, accessible to language, the broadest answer to this question is, probably, “most of it.” Even aspects of emotion or cognition not readily verbalized may be influenced by linguistically based thought processes. Thus, this conception is too broad to be of much use. (HCF, p. 1570)

It is hard to see this as anything other than a claim that approaches to linguistics focusing on anything that could fall under the label ‘E-language’ are to be dismissed as useless.

Some Externalists and Emergentists actually reject the idea that the human capacity for language yields “a potentially infinite array of expressions”. It is often pointed out by empirically inclined computational linguists that in practice there will only ever be a finite number of sentences to be dealt with (though the people saying this may underestimate the sheer vastness of the finite set involved). And naturally, for those who do not believe there are generative grammars in speakers’ heads at all, it holds a fortiori that speakers do not have grammars in their heads generating infinite languages (see Nefdt 2019c for a scientific modeling perspective on the infinity postulate). Externalists and Emergentists tend to hold that the “discrete infinity” that HCF posits is more plausibly a property of the generative Essentialists’ model of linguistic competence, I-language, or FLN, than a part of the human mind/brain. This does not mean that non-Essentialists deny that actual language use is creative, or (of course) that they think there is a longest sentence of English. But they may reject the link between linguistic productivity or creativity and the mathematical notion of recursion (see Pullum and Scholz 2010).

HCF’s remarks about how FLN “yields” or “generates” a specific “array” assume that languages are clearly and sharply individuated by their generators. They appear to be committed to the view that there is a fact of the matter about exactly which generator is in a given speaker’s head. Emergentists tend not to individuate languages in this way, and may reject generative grammars entirely as inappropriately or unacceptably ‘formalist’. They are content with the notion that the common-sense concept of a language is vague, and it is not the job of linguistic theory to explain what a language is, any more than it is the job of physicists to explain what material is, or of biologists to explain what life is. Emergentists, in particular, are interested not so much in identifying generators, or individuating languages, but in exploring the component capacities that facilitate linguistic communication, and finding out how they interact.

Similarly, Externalists are interested in the linguistic structure of expressions, but have little use for the idea of a discrete infinity of them, a view that is not, and cannot be empirically supported, unless one thinks of simplicity and elegance of theory as empirical matters. They focus on the outward manifestations of language, not on a set of expressions regarded as a whole language—at least not in any way that would give a language a definite cardinality. Zellig Harris, an archetypal Externalist, is explicit that the reason for not regarding the set of utterances as finite concerns the elegance of the resulting grammar: “If we were to insist on a finite language, we would have to include in our grammar several highly arbitrary and numerical conditions” (Harris 1957: 208). Infinitude, on his view is an unimportant side consequence of setting up a sentence-generating grammar in an uncluttered and maximally elegant way, not a discovered property of languages (see Pullum and Scholz 2010 for further discussion).

Not all Essentialists agree that linguistics studies aspects of what is in the mind or aspects of what is human. There are some who do not see language as either mental or human, and certainly do not regard linguists as working on a problem within cognitive psychology or neurophysiology. The debate on the ontology of language has seen three major options emerging in the literature. Besides the mentalism of Chomskyan linguistics, Katz (1981), Katz and Postal (1991) and Postal (2003) proffered a platonistic alternative and finally nominalism was proposed by Devitt (2006).

However, the Katzian trichotomy is no longer a useful characterisation of the state-of-the-art in linguistic ontology. For one thing, Katzian-style linguistic Platonism has very few if any extant adherents. One reason for this situation is that linguistic platonists attempt to restage the debate on the foundations and metaphysics of natural language within the philosophy of mathematics (see Katz 1996). But even if this move was legitimate, it would only have opened up a range of possibilities including nominalism (Field 1980; Azzouni 2004), structuralism (Hellman 1989; Shapiro 2007; Nefdt 2016), and forms of mentalism in the guise of intuitionism. For instance, while Richard Montague is often attributed with the view that linguistics can be viewed as a branch of mathematics, it is unclear whether or not he endorsed a platonistic ontology. Devitt (2006: 26) describes the possibility of a ‘methodological platonism’ in the following manner:

It is often convenient to talk of objects posited by these theories as if they were types not tokens, as if they were Platonic objects, but this need be nothing more than a manner of speaking: when the chips are down the objects are part of the spatiotemporal physical world.

Devitt’s nominalism or ‘linguistic conception’ was not around at the time of the original Katzian tripartite analysis. He argues that linguistics is an empirical science which studies languages as they are spoken by linguistic communities and viewing sentences as ‘idealised tokens’. Devitt’s ‘linguistic view’ (as opposed to the ’psychological view’ or Chomskyan mentalism) claims that grammars map onto behavioural output of language production, of which speakers are generally ignorant.

Katz took nominalism to have been refuted by Chomsky in his critiques of American structuralists in the 1960s. But, in Katz’s opinion, Chomsky had failed to notice that conceptualism was infected with many of the same faults as nominalism, because it too localized language spatiotemporally (in contingently existing, finite, human brains). Since contemporary Minimalist theories share in the earlier ontological commitment, Katz’ argument would presumably extent to them. Through an argument by elimination, Katz concluded that only platonism remained, and must be the correct view to adopt. But this is a false trichotomy and besides predating Devitt’s more philosophically grounded nominalism, it also fails to take linguistic pluralism into account.

Recent adherents of pluralism are Stainton (2014) and Santana (2016). Santana (2016) argues in favour of a pluralistic ontology for natural language based on all of the major foundational approaches, including sociolinguistic ontology. His approach is thoroughly naturalistic in asking the ontological question through the lens of “what sort of roles the concept of language plays in linguistic theory and practice” (Santana, 2016: 501).

The first thing Santana does is to separate the discussion into two related questions, one scientific and the other metascientific or ‘descriptive’ and ‘normative’ in his terms. He claims that “[l]anguage, the scientific concept, is thus descriptively whatever it is that linguists take as their primary object of study, and normatively whatever it is they should be studying” (Santana, 2016: 501). Eventually he advocates a union of various ontologies based on the ineliminable status of each perspective (in that way the oppose of Katz’ eliminative strategy).

Stainton (2014) similarly proposes a pluralistic ontology but with a more intersectional approach. His additional argument relates to how all of the views are indeed compatible. This argument is a response to an immediate objection along the lines of Postal (2003, 2009) as to the incompatibility of the various ontologies associated with mentalism, Platonism, physicalism and public language views. Stainton begins the pluralist apology in this way.

There is an obvious rebuttal on behalf of pluralism, namely that “the linguistic” is a complex phenomenon with parts that belong to distinct ontological categories. This shouldn’t surprise, since even “the mathematical” is like this: Two wholly physical dogs plus two other wholly physical dogs yields four dogs; there certainly is the mental operation of multiplying 26 by 84, the mental state of thinking about the square root of 7, and so on. (2014: 5)

His main argument against incompatibility, and in favour of intersection, is that the former rests on an equivocation of the terms ‘mental’, ‘abstract’ and even ‘physical’. Once the equivocation is cleared up, it is argued, hybrid ontological objects are licensed. The argument goes that appreciating the nuanced physical and mental and what he calls ‘abstractish’ nature of natural language will dissolve worries about ontological inconsistency and open the door for intersection. Consider some other members of this category of objects.

Indeed, our world is replete with such hybrid objects: psychocultural kinds (e.g. dining room tables, footwear, bonfires, people, sport fishing [...]; intellectual artifacts (college diplomas, drivers’ licenses, the Canadian dollar [...]; and institutions (MIT’s Department of Linguistics and Philosophy, Disneyworld [...] (Stainton, 2014: 6).

Despite the decline in interest in the ontology of language itself, philosophers have recently embraced a subset of this debate in the philosophy of linguistic objects with a special focus on words. There is a recent debate in philosophy on the philosophy of what Rey (2006, 2020) calls ‘Standard Linguistic Entities’ (SLEs) or tokens of word, sentence, morpheme, and phonemes types. Rey then defines a position called ’physical tokenism’ or PT as the assumption that SLEs can be identified with physical (acoustic) spatio-temporal phenomena. He doesn’t think that SLEs share the same kind of existence as his trusty Honda. In fact he thinks that they are ‘intentional inexistents’ (borrowed from Brentano, see the SEP entry on ‘Brentano’s Theory of Judgement, see also the section entitled ‘Intentional Inexistence’ in the entry on Intentionality ) or purely intentional uses of the term ‘represents’ which denote fictions of a particular sort. Linguistic theory according to him is only committed to the intentional contents of things like nouns, verbs, verb phrases etc. where “an (intentional) content is whatever we understand x to be when we use the idiom ‘represent(ation of) x’ but there is no real x” (2006: 242).

There has been some theoretical work on the nature of entities like phrases and words in linguistics. For example, Ross (2010) argues that the concept of parts of speech is fuzzy. Similarly, Szabó (2015) rejects the idea that parts of speech should be identified by distributional analysis as is common in syntax. Instead he offers a semantic approach based on predicate logic where the aim is to model the major lexical categories directly in terms of open class constants. This, he claims, results in a reduction of the gap between grammar and logic. So, for instance, nouns become not types corresponding to distributionally defined syntactic objects but rather open lexical constants used for reference such that the semantic clause only needs to involve a universal quantifier and a variable specified in terms of reference. Verbs, on the other hand, are constants which purport to predicate (for more details, see Szabó 2015 and Nefdt 2020). For Haspelmath (2011) a central problem is the concept of wordhood. He identifies ten morphosyntactic criteria for words as the best possible candidates over seemingly inferior semantic or phonological options. He shows all of them to be wanting, with the result among other things being that “the notion of lexical integrity is not well supported and should not be appealed to in explaining grammatical phenomena” (Haspelmath, 2011: 33). The very notion of wordhood, although intuitive and central, is unclear upon further scrutiny. Yet, in linguistics there is continual hope for a resolution, that there is something more than essential inexistence at stake. Haspelmath thinks this is a vain hope, and attributes it to the influence of orthography on the thought of linguist researchers.

Philosophers have been traditionally interested in the metaphysics of SLEs with a special focus on the ontological status of words. Interestingly, this literature showcases variations on the foundational debates on the ontology of language. As Miller (2020) notes:

Words play various roles in our lives. Some insult, some inspire, and words are central to communication. The aim of an ontology of words is to determine what entities, if any, can play those roles and possess (or instantiate) these properties. (2)

However the positions advocated are somewhat more nuanced than the original Katzian trichotomy suggests. They usually start with the problem of word individuation expressed in the following manner:

Think of the following line: A rose is a rose is a rose. How many words are there in this line? If we were to count words themselves, not their instances, the answer is three: rose, is, and a. If we were to count the concrete instances we see on a piece of paper, the answer is eight. The line, however, can be taken as an abstract type; a sequence of shapes. (Irmak, 2019: 1140)

Further complications are introduced by reference to words like “color” and “colour”. Here the idea is that the phonological profile of a word is a guide to its identity. But this fails in other cases.

Now take the name MOHAMMED. Since Arabic does not notate vowels, the name has been transcribed in a wide variety of ways in English, and some of such transcriptions present important discrepancies: For example, “Mohammad” and “Mehmood.” Even if we know that they originated from the same source, the difference between the two forms is considerable, and intuitions about their being variants of the same name are less clear. What is the point up to which differences in spelling are consistent with word identity? (Gasparri, 2021: 594)

Word individuation goes beyond this initial characterisation and it is not always clear how the many accounts deal with the more complex questions directly in their metaphysical pursuits. For instance, issues not usually mentioned in the literature, but which seem equally important are related to whether pitch in ‘They poured pitch all over the parking lot’ and ‘The players swarmed the pitch after they won the game’ are different words. What about words within different syntactic categories, such as ‘watch’ (time-telling device) and ‘watch’ (observe)? Are “ain’t” and “isn’t” and “aren’t” different words? What about simple cases of inflectional morphology such as like ‘toy’ and ‘toys’?

According to Nefdt (2019b), the identity of a word is tied to its role in the sentence structure. In which case, “ain’t” and “isn’t” come out as the same word (at least in the singular use) but ‘watch’ (noun) and ‘watch’ (verb) do not. However, counterintuitively, his account might license the identity of words like ‘truck’ and ‘lorry’.

There are two strong but separate traditions which can both lay claim to being the ‘received position’. Within linguistics, the idea of a word as a LEXEME or mental dictionary entry is commonplace (with stipulations for senses, irregular forms, and selectional criteria). Most introductory textbooks assume something of this sort. In the philosophical literature, on the other hand, a mild or methodological version of platonism is often presupposed. This view has it that words can be separated into types and tokens, where the former lack specific spatiotemporal features and the latter instantiate these forms somehow.

The latter intuition seems to characterize most views on the ontology of words. Bromberger (1989) defines what he calls “the Platonic Relationship Principle” or the principle that allows us to ‘impute properties to types after observing and judging some of their tokens’ (Bromberger 1989, 62). While Bromberger (1989, 2011) represents the pinnacle of the classical philosophy of linguistics approach to these questions. In a more metaphysical mode, David Kaplan (1990, 2011) constructs a thoroughly physicalist proposal in which words are modelled in terms of a stages and continuants:

I propose a quite different model according to which utterances and inscriptions are stages of words, which are the continuants made up of these interpersonal stages along with some more mysterious intrapersonal stages. (Kaplan 1990: 98)

For him, what individuates words is the intention of the user (see Cappelen (1999) for an objection to intentional accounts tout court). Unfortunately, fascinating as Kaplan’s proposal is, it does not attempt to reflect on linguistic theory directly. In fact, one major criticism of his view, courtesy of Hawthorne and Lepore (2011), is that it fails to account for uninstantiated word-types whose existence is guaranteed by derivational morphology, whether or not they’ve been tokened or baptized in the real world. Other notable accounts are Wetzel’s (2009) Platonism (see the SEP entry on abstract objects ) and Szabó’s (1999) representational/nominalist view.

The philosophy of words has recently seen a resurgence in interest among philosophers, especially on the ontological issues. Miller (2021), for example, attempts to apply a bundle theory to the task of word individuation and identification. Irmak (2019) suggests that words are abstract artifacts (similarly to Katz and Wetzel) but insists that they are more akin to musical scores or works of fiction which have temporal components (“temporal abstracta” as he calls it). Mallory (2020) advocates the position that words are not really objects in the ordinary sense. However, he opts for an action-theoretic approach in which tokens provide instructions for the performance of action-types where our normal understanding of ‘word’ is to be identified with those types. His view is overtly naturalistic and focuses on the concept of words which is drawn from contemporary linguistic theory. Similarly, Nefdt (2019b) proffers a mathematical structuralist interpretation of SLEs in which the definition of words is continuous with the ontology of phrases and sentences. Here he follows Jackendoff (2018) who uses model-theoretic (Pullum 2013) or constraint-based grammar formalisms to argue for a continuum between words and linguistic rules. In other words, these latter two authors reject the idea that words are somehow sui generis entities in need of discontinuous explanation. Gasparri (2020) suggests pluralism is a more solid foundation for the ontology of words. He evaluates both “bundlistic” views such as Miller’s and causal-historical accounts such as Irmak’s before offering an alternative “view that there is a plurality of epistemically virtuous ways of thinking about the nature of words” (608).

These are of course complex issues and they offer a lens through which to appreciate the erstwhile debate on the ontology of language but with a contemporary and more focused flavor. Not all of the authors who work on the philosophy of words consider the role of linguistic theory to be central. Hence their work might be related but it does not quite qualify as the philosophy of linguistics, where this is viewed as a subfield of the philosophy of science. By contrast, we have focused on the authors who directly engage with linguistic theory in their accounts of the ontology of SLEs. There is also no clear mapping between the various ontological accounts mentioned here and the characterizations of linguistic theorizing in terms of Externalism, Emergentism and Essentialism. No particular metaphysical view unifies any of our three groupings. For example, not all Externalists incline toward nominalism; numerous Emergentists as well as most Essentialists take linguistics to be about mental phenomena; and our Essentialists include Katz’s platonism alongside the Chomskyan ‘I-language’ advocates and pluralists embrace aspects of all of the above.

Linguists’ conception of the components of the study of language contrast with philosophers’ conceptions (even those of philosophers of language) in at least three ways. First, linguists are often intensely interested in small details of linguistic form in their own right. Second, linguists take an interest in whole topic areas like the internal structure of phrases, the physics of pronunciation, morphological features such as conjugation classes, lexical information about particular words, and so on—topics in which there is typically little philosophical payoff. And third, linguists are concerned with relations between the different subsystems of languages: the exact way the syntax meshes with the semantics, the relationship between phonological and syntactic facts, and so on.

With regard to form, philosophers broadly follow Morris (1938), a foundational work in semiotics, and to some extent Peirce (see SEP entry: Peirce, semiotics), in thinking of the theory of language as having three main components:

  • syntax , which treats of the form of signs;
  • semantics , which deals with the relations of signs to their denotations; and
  • pragmatics , which concerns the contextualized use of interpreted signs.

Linguists, by contrast, following both Sapir (1921) and Bloomfield (1933), treat the syntactic component in a more detailed way than Morris or Peirce, and distinguish between at least three kinds of linguistic form: the form of speech sounds (phonology), the form of words (morphology), and the form of sentences. (If syntax is about the form of expressions in general, then each of these would be an element of Morris’s syntax.)

Emergentists in general deny that there is a distinction between semantics and pragmatics—a position that is familiar enough in philosophy: Quine (1987: 211), for instance, holds that “the separation between semantics and pragmatics is a pernicious error.” And generally speaking, those theorists who, like the later Wittgenstein, focus on meaning as use will deny that one can separate semantics from pragmatics. Emergentists such as Paul Hopper & Sandra Thompson agree:

[W]hat is called semantics and what is called pragmatics are an integrated whole. (Hopper and Thompson 1993: 372)

Some Essentialists—notably Chomsky—also deny that semantics can be separated from pragmatics, but unlike the Emergentists (who think that semantics-pragmatics is a starting point for linguistic theory), Chomsky (as we noted briefly in section 1.3) denies that semantics and pragmatics can have any role in linguistics:

It seems that other cognitive systems—in particular, our system of beliefs concerning things in the world and their behavior—play an essential part in our judgments of meaning and reference, in an extremely intricate manner, and it is not at all clear that much will remain if we try to separate the purely linguistic components of what in informal usage or even in technical discussion we call ‘the meaning of [a] linguistic expression.’ (Chomsky 1979; 142)

Regarding the theoretical account of the relation between words or phrases and what speakers take them to refer to, Chomsky says, “I think such theories should be regarded as a variety of syntax” (Chomsky 1992: 223).

Not every Essentialist agrees with Chomsky on this point. Many believe that every theory should incorporate a linguistic component that yields meanings, in much the same way that many philosophers of language believe there to be such a separate component. Often, although not always, this component amounts to a truth-theoretic account of the values of syntactically-characterized sentences. This typically involves a translation of the natural language sentence into some representation that is “intermediate” between natural language and a truth-theory—perhaps an augmented version of first-order logic, or perhaps a higher-order intensional language. The Essentialists who study semantics in such ways usually agree with Chomsky in seeing little role for pragmatics within linguistic theory. But their separation of semantics from pragmatics allows them to accord semantics a legitimacy within linguistics itself, and not just in psychology or sociology.

Such Essentialists, as well as the Emergentists, differ in important ways from classical philosophical logic in their attitudes towards “the syntactic-semantic interface”, however. Philosophers of language and logic who are not also heavily influenced by linguistics tend to move directly—perhaps by means of a “semantic intuition” or perhaps from an intuitive understanding of the truth conditions involved—from a natural language sentence to its “deep, logical” representation. For example, they may move directly from (EX1) to (LF1):

And from there perhaps to a model-theoretic description of its truth-conditions. A linguist, on the other hand, would aim to describe how (EX1) and (LF1) are related. From the point of view of a semantically-inclined Essentialist, the question is: how should the syntactic component of linguistic theory be written so that the semantic value (or, “logical form representation”) can be assigned? From some Emergentist points of view, the question is: how can the semantic properties and communicative function of an expression explain its syntactic properties?

Matters are perhaps less clear with the Externalists—at least with those who identify semantic value with distribution in terms of neighboring words (there is a tradition stemming from the structuralists of equating synonymy with the possibility of substitution in all contexts without affecting acceptability).

Matters are in general quite a bit more subtle and tricky than (EX1) might suggest. Philosophers have taken the natural language sentence (EX2) to have two logical forms, (LF2a) and (LF2b):

But for the linguist interested in the syntax-semantics interface, there needs to be some explanation of how (LF2a) and (LF2b) are associated with (EX2). It could be a way in which rules can derive (LF2a) and (LF2b) from the syntactic representation of (EX2), as some semantically-inclined Essentialists would propose, or a way to explain the syntactic properties of (EX2) from facts about the meanings represented by (LF2a) and (LF2b), as some Emergentists might want. But that they should be connected up in some way is something that linguists would typically count as non-negotiable.

3. Linguistic Methodology and Data

The strengths and limitations of different data gathering methods began to play an important role in linguistics in the early to mid-20th century. Voegelin and Harris (1951: 323) discuss several methods that had been used to distinguish Amerindian languages and dialects:

  • Informal elicitation : asking an informant for a metalinguistic judgment on an expression. [E.g., “Is this sentence grammatical?” “Do these two sentences mean the same thing?”]
  • Corpus collection : gathering a body of naturally occurring utterances.
  • Controlled experimentation : testing informants in some way that directly gauges their linguistic capacities.

They note that the anthropological linguists Boas and Sapir (who we take to be proto-Emergentists) used the ‘ask the informant’ method of informal elicitation, addressing questions “to the informant’s perception rather than to the data directly” (1951: 324). Bloomfield (the proto-Externalist), on the other hand, worked on Amerindian languages mostly by collecting corpora, with occasional use of monolingual elicitation.

The preferred method of Essentialists today is informal elicitation, including elicitation from oneself. Although the techniques for gathering data about speakers and their language use have changed dramatically over the past 60 or more years, the general strategies have not: data is still gathered by elicitation of metalinguistic judgments, collection of corpus material, or direct psychological testing of speakers’ reactions and behaviors. Different linguists will have different preferences among these techniques, but it is important to understand that data could be gathered in any of the three ways by advocates of any tendency. Essentialists, Emergentists, and Externalists differ as much on how data is interpreted and used as on their views of how it should be gathered.

A wide range of methodological issues about data collection have been raised in linguistics. Since gathering data by direct objective experimental testing of informants is a familiar practice throughout the social, psychological, medical, and biological sciences, we will say little about it here, focusing instead on these five issues about data:

  • Disputes over the use of linguistic intuitions as linguistic data;
  • Differences between grammaticality and acceptability judgments;
  • Differences between scales for measuring acceptability judgments;
  • Debates about the reliability of informal judgment elicitation methods; and
  • Issues concerning the relevance and reliability of corpus evidence.

The debate in linguistics over the use of linguistic intuitions (elicited metalinguistic judgments) as data, and how that data should be collected has resulted in enduring, rancorous, often ideologically tinged disputes over the past 45 years. The disputes are remarkable, if only for their fairly consistent venomous tone.

At their most extreme, many Emergentists and some Externalists cast the debate in terms of whether linguistic intuitions should ever count as evidence for linguistic theorizing. And many Essentialists cast it in terms of whether anything but linguistic intuitions are ever really needed to support linguistic theorizing.

The debate focuses on the Essentialists’ notion of a mental grammar, since linguistic intuitions are generally understood to be a consequence of tacit knowledge of language. Emergentists who deny that speakers have innate domain-specific grammars (competence, I-languages, or FLN) have raised a diverse range of objections to the use of reports of intuitions as linguistic data – though Devitt (2006) offers an account of linguistic intuitions that does not base them on inferred tacit knowledge of competence grammars. The following passages are representative Emergentist critiques of ‘intuitions’ (elicited judgments):

Generative linguists typically respond to calls for evidence for the reality of their theoretical constructs by claiming that no evidence is needed over and above the theory’s ability to account for patterns of grammaticality judgments elicited from native speakers. This response is unsatisfactory on two accounts. First, such judgments are inherently unreliable because of their unavoidable meta-cognitive overtones… Second, the outcome of a judgment (or the analysis of an elicited utterance) is invariably brought to bear on some distinction between variants of the current generative theory, never on its foundational assumptions. (Edelman and Christiansen 2003: 60) The data that are actually used toward this end in Generative Grammar analyses are almost always disembodied sentences that analysts have made up ad hoc, … rather than utterances produced by real people in real discourse situations… In diametric opposition to these methodological assumptions and choices, cognitive-functional linguists take as their object of study all aspects of natural language understanding and use… They (especially the more functionally oriented analysts) take as an important part of their data not disembodied sentences derived from introspection, but rather utterances or other longer sequences from naturally occurring discourse. (Tomasello 1998: xiii) [T]he journals are full of papers containing highly questionable data, as readers can verify simply by perusing the examples in nearly any syntax article about a familiar language. (Wasow and Arnold 2005: 1484)

It is a common Emergentist objection that linguistic intuitions (taken to be reports of elicited judgments of the acceptability of expressions not their grammaticality) are bad data points because not only are they not usage data, i.e., they are metalinguistic, but also because they are linguists’ judgments about invented example sentences. On neither count would they be clear and direct evidence of language use and human communicative capacities—the subject matter of linguistics on the Emergentist view. A further objection is to their use by theorists to the exclusion of all other kinds of evidence. For example,

[Formal linguistics] continues to insist that its method for gathering data is not only appropriate, but is superior to others. Occasionally a syntactician will acknowledge that no one type of data is privileged, but the actual behavior of people in the field belies this concession. Take a look at any recent article on formal syntax and see whether anything other than the theorist’s judgments constitute the data on which the arguments are based. (Ferreira 2005: 372)

“Formal” is Emergentist shorthand for referring to generative linguistics. And it should be noted that the practice by Essentialists of collapsing various kinds of acceptability judgments under the single label ‘intuitions’ masks important differences. In principle there might be significant differences between the judgments of (i) linguists with a stake in what the evidence shows; (ii) linguists with experience in syntactic theory but no stake in the issue at hand; (iii) non-linguist native speakers who have been tutored in how to provide the kinds of judgments the linguist is interested in; and (iv) linguistically naïve native speakers.

Many Emergentists object to all four kinds of reports of intuitions on the grounds that they are not direct evidence language use. For example, a common objection is based on the view that

[T]he primary object of study is the language people actually produce and understand. Language in use is the best evidence we have for determining the nature and specific organization of linguistic systems. Thus, an ideal usage-based analysis is one that emerges from observation of such bodies of usage data, called corpora.… Because the linguistic system is so closely tied to usage, it follows that theories of language should be grounded in an observation of data from actual uses of language. (Barlow and Kemmer 2002, Introduction)

But collections of linguists’ reports of their own judgments are also criticized by Emergentists as “arm-chair data collection,” or “data collection by introspection”. All parties tend to call this kind of data collection “informal”—though they all rely on either formally or informally elicited judgments to some degree.

On the other side, Essentialists tend to deny that usage data is adequate evidence by itself:

More than five decades of research in generative linguistics have shown that the standard generative methodology of hypothesis formation and empirical verification via judgment elicitation can lead to a veritable goldmine of linguistic discovery and explanation. In many cases it has yielded good, replicable results, ones that could not as easily have been obtained by using other data-gathering methods, such as corpus-based research…[C]onsider the fact that parasitic gap constructions…are exceedingly rare in corpora…. [T]hese distributional phenomena would have been entirely impossible to distill via any non-introspective, non-elicitation based data gathering method. Corpus data simply cannot yield such a detailed picture of what is licit and, more crucially, what is not licit for a particular construction in a particular linguistic environment. (den Dikken et al. 2007: 336)

And Essentialists often seem to deny that they are guilty of what the Emergentist claims they are guilty of. For example, Chomsky appears to be claiming that acceptability judgments are performance data, i.e. evidence of use:

Acceptability is a concept that belongs to the study of performance, whereas grammaticalness belongs to the study of competence… Like acceptability, grammaticalness is, no doubt, a matter of degree…but the scales of grammaticalness and acceptability do not coincide. Grammaticalness is only one of many factors that interact to determine acceptability. (Chomsky 1965: 11)

Chomsky means to deny that acceptability judgments are direct evidence of linguistic competence . But it does not follow from this that elicited acceptability judgments are direct evidence of language use.

And as for the charge of “arm-chair” collection methods, some Essentialists claim to have shown that such methods are as good as more controlled experimental methods. For example, Sprouse and Almeida report:

[W]e empirically assess this claim by formally testing all 469 (unique, US-English) data points from a popular syntax textbook (Adger 2003) using 440 naïve participants, two judgment tasks (magnitude estimation and yes–no), and three different types of statistical analyses (standard frequentist tests, linear mixed effects models, and Bayes factor analysis). The results suggest that the maximum discrepancy between traditional methods and formal experimental methods is 2%. This suggests that … the minimum replication rate of these 469 data points is 98%. (Spouse and Almeida 2012, p. 609, abstract)

This can be read as defending either Essentialists’ consulting of their own intuitions simpliciter, or their self-consultation of intuitions on uncontroversial textbook cases only. The former is much more controversial than the later.

One might also wonder whether an error rate of 2% really is appropriate for the primary data presented in an elementary textbook. If a geography textbook misidentified 2–3% of the rivers of the continental United States, or gave incorrect locations for them, or incorrectly reported their lengths, it would forfeit our trust. Analogous claims could be made about any elementary textbook in other fields: an elementary English literature textbook that misidentified the authors of 2% of the books discussed, or their years of publication, etc.

Finally, both parties of the debate engage in ad hominem attacks on their opponents. Here is one example of a classic ad hominem or tu quoque attack on Emergentists in defense of constructed examples by Essentialists:

[The charge made concerning “armchair data collection”] implies that there is something intrinsic to generative grammar that invites partisans of that framework to construct syntactic theories on the evidence of a single person’s judgments. Nothing could be farther from the truth. The great bulk of publications in cognitive and functional linguistics follow the same practice. Of course, rhetorically many of the latter decry the use of linguists’ own intuitions as data. For example, in … an important collections [sic] of papers in cognitive-functional linguistics, … only two contributors to the volume … present segments of natural discourse, neither filling even a page of text. All of the other contributors employ examples constructed by the linguists themselves. It is quite difficult to find any work in cognitive linguistics (and functional linguists are only slightly better) that uses multiple informants. It seems almost disingenuous … to fault generativists for what (for better or worse) is standard practice in the field, regardless of theoretical allegiance. (Newmeyer 2007: 395)

Clearly, the mere fact that some Emergentists may in practice have made use of invented examples in testing their theories does not tell against any cogent general objections they may have offered to such practice. What is needed is a decision on the methodological point, not just a cry of “You did it too!”.

Given the intolerance of each other’s views, and the crosstalk present in these debates, it is tempting to think that Emergentism and Essentialism are fundamentally incompatible on what counts as linguistic data, since their differences are based on their different views of the subject matter of linguistics, and what the phenomena and goals of linguistic theorizing are. There is no doubt that the opposing sides think that their respective views are incompatible. But this conclusion may well be too hasty. In what follows, we try to point to a way that the dispute could be ameliorated, if not adjudicated.

Essentialists who accept the competence/performance distinction of Chomsky (1965) traditionally emphasize elicited acceptability judgment data (although they need not reject data that is gathered using other methods). But as Cowart notes:

In this view, which exploits the distinction between competence and performance, the act of expressing a judgment of acceptability is a kind of linguistic performance. The grammar that a [generative Essentialist] linguistic theory posits in the head of a speaker does not exercise exhaustive control of judgments… While forming a sentence judgment, a speaker draws on a variety of cognitive resources… The resulting [acceptability] judgments could pattern quite differently than the grammaticality values we might like them to reflect. (Cowart 1997: 7)

The grammaticality of an expression, on the standard generative Essentialist view, is the status conferred on it by the competence state of an ideal speaker. But competence can never be exercised or used without potentially interfering performance factors like memory being exercised as well. This means that judgments about grammaticality are never really directly available to the linguist through informant judgments: they have to be inferred from judgments of acceptability (along with any other relevant evidence). Nevertheless, Essentialists do take acceptability judgments to provide fairly good evidence concerning the character of linguistic competence. In fact the use of informally gathered acceptability judgment data is a hallmark of post-1965 Essentialist practice.

It would be a mistake, however, to suppose that only Essentialists make use of such judgments. Many contemporary Externalists and Emergentists who reject the competence/performance distinction still use informally gathered acceptability judgments in linguistic theorizing, though perhaps not in theory testing. Emergentists tend to interpret experimentally gathered judgment data as performance data reflecting the interactions between learned features of communication systems and general learning mechanisms as deployed in communication. And Externalists use judgment data for corpus cleaning (see below).

It should be noted that sociolinguists and anthropological linguists (and we regard them as tending toward Emergentist views) often informally elicit informant judgments not only about acceptability but also about social and regional style and variation, and meaning. They may ask informants questions like, “Who would typically say that?”, or “What does X mean in context XYZ?”, or “If you can say WXY, can you say WXZ?” (see Labov 1996: 77).

A generative grammar gives a finite specification of a set of expressions. A psychogrammar, to the extent that it corresponds to a generative grammar, might be thought to equip a speaker to know (at least in principle) absolutely whether a string is in the language. However, elicited metalinguistic judgments are uncontroversially a matter of degree. A question arises concerning the scale on which these degrees of acceptability should be measured.

Linguists have implicitly worked with a scale of roughly half a dozen levels and types of acceptability, annotating them with prefixed symbols. The most familiar is the asterisk, originally used simply to mark strings of words as ungrammatical, i.e., as not belonging to the language at all. Other prefixed marks have gradually become current:

But other annotations have been used to indicate a gradation in the extent to which some sentences are unacceptable. No scientifically validated or explicitly agreed meanings have been associated with these marks, but a tradition has slowly grown up of assigning prefixes such as those in Table 2 to signify degrees of unacceptability:

Table 2: Prefixes used to mark levels of acceptability

Such markings are often used in a way that suggests an ordinal scale , i.e. a partial ordering that is silent on anything other than equivalence in acceptability or ranking in degree of unacceptability.

By contrast, Bard et al. (1996: 39) point out, it is possible to use interval scales , which additionally measure distance between ordinal positions. Interval scales of acceptability would measure relative distances between strings—how much more or less acceptable one is than another. Magnitude estimation is a method developed in psychophysics to measure subjects’ judgments of physical stimuli on an interval scale. Bard et al. (1996) adapted these methods to linguistic acceptability judgments, arguing that interval scales of measurement are required for testing theoretical claims that rely on subtle judgments of comparative acceptability. An ordinal scale of acceptability can represent one expression as being less acceptable than another, but cannot support quantitative questions about how much less. Many generative Essentialist theorists had been suggesting that violation of different universal principles led to different degrees of unacceptability. According to Bard et al. (34–35), because there may be “disproportion between the fineness of judgments people can make and the symbol set available for recording them” it will not suffice to use some fixed scale such as this one:

? < ?? < ?* < * < **

indicating absolute degrees of unacceptability. Degrees of relative unacceptability must be measured. This is done by asking the informant how much less acceptable one string is than another.

Magnitude estimation can be used with both informal and experimental methods of data collection. And data that is measured using interval scales can be subjected to much more mathematically sophisticated tests and analyses than data measured solely by an ordinal scale, provided that quantitative data are available.

It should be noted that the value of applying magnitude estimation to the judgment of acceptability has been directly challenged in two recent papers. Weskott and Fanselow (2011) and Sprouse (2011) both present critiques of Bard et al. (1996). Weskott and Fanselow compared magnitude estimation data to standard judgments on binary and 7-point scales, and claim that magnitude estimation does not yield more information than other judgment tasks, and moreover can produce spurious variance. And Sprouse, on the basis of recent formalizations of magnitude estimation in the psychophysics literature, presents experimental evidence that participants cannot make ratio judgments of acceptability (for example, a judgment that one sentence is precisely half as acceptable as another), which suggests that the magnitude estimation task probably provides the same interval-level data as other judgment tasks.

Part of the dispute over the reliability of informal methods of acceptability judgment elicitation and collection is between different groups of Essentialists. Experimentally trained psycholinguists advocate using and adapting various experimental methods that have been developed in the cognitive and behavioral sciences to collect acceptability judgments. And while the debate is often cast in terms of which method is absolutely better, a more appropriate question might be when one method is to be preferred to the others. Those inclined toward less experimentally controlled methods point out that there are many clear and uncontroversial acceptability judgments that do not need to be shown to be reliable. Advocates of experimental methods point out that many purportedly clear, uncontroversial judgments have turned out to be unreliable, and led to false empirical generalizations about languages. Both seem to be right in different cases.

Chomsky has frequently stated his view that the experimental data-gathering techniques developed in the behavioral sciences are neither used nor needed in linguistic theorizing. For example:

The gathering of data is informal; there has been little use of experimental approaches (outside of phonetics) or of complex techniques of data collection and data analysis of a sort that can easily be devised, and that are widely used in the behavioral sciences. The arguments in favor of this informal procedure seem to me quite compelling; basically, they turn on the realization that for the theoretical problems that seem most critical today, it is not at all difficult to obtain a mass of crucial data without use of such techniques. Consequently, linguistic work, at what I believe to be its best, lacks many of the features of the behavioral sciences. (Chomsky 1969: 56)

He also expressed the opinion that using experimental behavioral data collection methods in linguistics “would be a waste of time and energy” (1969: 81).

Although many Emergentists—the intellectual heirs of Sapir—would accept ‘ask-the-informant’ data, we might expect them to tend to accept experimental data-gathering methods that have been developed in the social sciences. There is little doubt that strict followers of the methodology preferred by Bloomfield in his later career would disapprove of ‘ask the informant’ methods. Charles Hockett remarked:

A language, as a set of habits, is a fragile thing, subject to minor modification in the slightest breeze of circumstance; this, indeed, is its great source of power. But this is also why the transformationalists (like the rest of us!), using themselves as informants, have such a hard time deciding whether certain candidates for sentencehood are really ‘in their dialect’ or not; and it is why Bloomfield, in his field work, would never elicit paradigms, for fear he would induce his informant to say something under the artificial conditions of talking with an outsider that he would never have said in his own everyday surroundings. (Hockett 1968: 89–90, fn. 31)

We might expect Bloomfield, having abandoned his earlier Wundtian psychological leanings, to be suspicious of any method that could be cast as introspective. And we might expect many contemporary Externalists to prefer more experimentally controlled methods too. (We shall see below that to some extent they do.)

Derwing (1973) was one early critic of Chomsky’s view (1969) that experimentally controlled data collection is useless; but it was nearly 25 years before systematic research into possible confounding variables in acceptability judgment data started being conducted on any significant scale. In the same year that Bard et al. (1996) appeared, Carson Schütze (1996) published a monograph with the following goal statement:

I aim to demonstrate…that grammaticality judgments and other sorts of linguistic intuition, while indispensable forms of data for linguistic theory, require new ways of being collected and used. A great deal is known about the instability and unreliability of judgments, but rather than propose that they be abandoned, I endeavor to explain the source of their shiftiness and how it can be minimized. (1996: 1)

In a similar vein, Wayne Cowart stated that he wanted to “describe a family of practical methods that yield demonstrably reliable data on patterns of sentence acceptability.” He observes that the stability and reliability of acceptability judgment collection is

complicated by the fact that there seems to be no consensus on how to gather judgments apart from a widespread tolerance for informal methods in which the linguist consults her own intuitions and those of the first handy informant (what we might call the “Hey, Sally” method). (Cowart 1997: 2)

Schütze also expresses the importance of using experimental methods developed in cognitive science:

[M]y claim is that none of the variables that confound metalinguistic data are peculiar to judgments about language. Rather they can be shown to operate in some other domain in a similar way. (This is quite similar to Valian’s (1982) claim that the data of more traditional psychological experiments have all the same problems that judgment data have.) (Schütze 1996: 14)

The above can be read as sympathetic to the Essentialist preference for elicited judgments.

Among the findings of Schütze and Cowart about informal judgment collection methods are these:

  • There is really no agreement in linguistics on what counts as an informal method (though note that Sprouse and Almeida 2012 are much more comfortable with the informal method of consulting one's own intuitions of grammaticality).
  • The collection of acceptability judgment data is just as vulnerable to the influence of extraneous variables as are other kinds of psychological data.
  • Judgment samples can be biased in informal judgment collection.
  • Experimenter bias is often not controlled for in informal judgment collection.
  • Judgment materials are often not carefully prepared to present a relevant, well-ordered, contrasting set of minimal pairs.
  • The instability of one-off speaker judgments can be controlled for by gathering judgments from a given speaker across time.

Although Schütze (1996) and Cowart (1997) are both critical of traditional Essentialist informal elicitation methods, their primary concern is to show how the claims of Essentialist linguistics can be made less vulnerable to legitimate complaints about informal data collection methods. Broadly speaking, they are friends of Essentialism. Critics of Essentialism have raised similar concerns in less friendly terms, but it is important to note that the debate over the reliability of informal methods is a debate within Essentialist linguistics as well.

Informal methods of acceptability judgment data have often been described as excessively casual. Ferreira described the informal method this way:

An example sentence that is predicted to be ungrammatical is contrasted with some other sentence that is supposed to be similar in all relevant ways; these two sentences constitute a “minimal pair”. The author of the article provides the judgment that the sentence hypothesized to be bad is in fact ungrammatical, as indicated by the star annotating the example. But there are serious problems with this methodology. The example that is tested could have idiosyncratic properties due to its unique lexical content. Occasionally a second or third minimal pair is provided, but no attempt is made to consider the range of relevant extraneous variables that must be accounted for and held constant to make sure there isn’t some correlated property that is responsible for the contrast in judgments. Even worse, the “subject” who provides the data is not a naïve informant, but is in fact the theorist himself or herself, and that person has a stake in whether the sentence is judged grammatical or ungrammatical. That is, the person’s theory would be falsified if the prediction were wrong, and this is a potential source of bias. (Ferreira 2005: 372)

(It would be appropriate to read ‘grammatical’ and ‘grammaticality’ in Ferreira’s text as meaning ‘acceptable’ and ‘acceptability’.)

This critical characterization exemplifies the kind of method that Schütze and Cowart aimed to improve on. More recently, Gibson and Fedorenko describe the traditional informal method this way:

As has often been noted in recent years (Cowart, 1997; Edelman & Christiansen, 2003; Featherston, 2007; Ferreira, 2005; Gibson & Fedorenko, 2010a; Marantz, 2005; Myers, 2009; Schütze, 1996; Wasow & Arnold, 2005), the results obtained using this method are not necessarily generalisable because of (a) the small number of experimental participants (typically one); (b) the small number of experimental stimuli (typically one); (c) cognitive biases on the part of the researcher and participants; and (d) the effect of the preceding context (e.g., other constructions the researcher may have been recently considering). (Gibson and Fedorenko, 2013)

While some Essentialists have acknowledged these problems with the reliability of informal methods, others have, in effect, denied their relevance. For example, Colin Phillips (2010) argues that “there is little evidence for the frequent claim that sloppy data-collection practices have harmed the development of linguistic theories”. He admits that not all is epistemologically well in syntactic theory, but adds, “I just don’t think that the problems will be solved by a few rating surveys.” He concludes:

I do not think that we should be fooled into thinking that informal judgment gathering is the root of the problem or that more formalized judgment collection will solve the problems. (Phillips 2010: 61)

To suggest that informal methods are as fully reliable as controlled experimental ones would be a serious charge, implying that researchers like Bard, Robinson, Sorace, Cowart, Schütze, Gibson, Fedorenko, and others have been wasting their time. But Phillips actually seems to be making a different claim. He suggests first that informally gathered data has not actually harmed linguistics, and second that linguists are in danger of being “fooled” by critics who invent stories about unreliable data having harmed linguistics.

The harm that Phillips claims has not occurred relates to the charge that “mainstream linguistics” (he means the current generative Essentialist framework called ‘Minimalism’) is “irrelevant” to broader interests in the cognitive sciences, and has lost “the initiative in language study”. Of course, Phillips is right in a sense: one cannot insure that experimental judgment collection methods will address every way in which Minimalist theorizing is irrelevant to particular endeavors (language description, language teaching, natural language processing, or broader questions in cognitive psychological research). But this claim does not bear on what Schütze (1996) and Cowart (1997) show about the unreliability of informal methods.

Phillips does not fully accept the view of Chomsky (1969) that experimental methods are useless for data gathering (he says, “I do not mean to argue that comprehensive data gathering studies of acceptability are worthless”). But his defense of informal methods of data collection rests on whether these methods have damaged Essentialist theory testing:

The critiques I have read present no evidence of the supposed damage that informal intuitions have caused, and among those who do provide specific examples it is rare to provide clear evidence of the supposed damage that informal intuitions have caused… What I am specifically questioning is whether informal (and occasionally careless) gathering of acceptability judgments has actually held back progress in linguistics, and whether more careful gathering of acceptability judgments will provide the key to future progress.

Either Phillips is fronting the surprising opinion that generative theorizing has never been led down the wrong track by demonstrably unreliable data, or he is changing the subject. And unless clear criteria are established for what counts as “damage” and “holding back,” Phillips is not offering any testable hypothesis about data collection methodology. For example, Phillips discounts the observation of Schütze (1996) that conflicting judgments of relative unacceptability of violations of two linguistic universals held back the development of Government and Binding (GB), on the grounds that two sets of conflicting judgments and their analyses “are now largely forgotten, supplanted by theories that have little to say about such examples.” But the fact that the proposed universals are discarded principles of UG is irrelevant to the effect that unreliable data once had on the (now largely abandoned) GB theory. A methodological concern cannot be dismissed on the basis of a move to a new theory that abandons the old theory but not its methods!

More recently, Bresnan (2007) claims that many theoretical claims have arguably been supported by unreliable informally gathered syntactic acceptability judgments. She observes:

Erroneous generalizations based on linguistic intuitions about isolated, constructed examples occur throughout all parts of the grammar. They often seriously underestimate the space of grammatical possibility (Taylor 1994, 1996, Bresnan & Nikitina 2003, Fellbaum 2005, Lødrup 2006, among others), reflect relative frequency instead of categorical grammaticality (Labov 1996, Lapata 1999, Manning 2003), overlook complex constraint interactions (Green 1971, Gries 2003) and processing effects (Arnon et al. 2005a, b), and fail to address the problems of investigator bias (Labov 1975, Naro 1980, Chambers 2003: 34) and social intervention (Labov 1996, Milroy 2001, Cornips & Poletto 2005). (Bresnan 2007: 301)

Her discussion supports the view that various highly abstract theoretical hypotheses have been defended through the use of generalizations based on unreliable data.

The debate over the harm that the acceptance of informally collected data has had on theory testing is somewhat difficult to understand for Essentialist, Externalist, and Emergentist researchers who have been trained in the methods of the cognitive and behavioral sciences. Why try to support one’s theories of universal grammar, or of the grammars of particular languages, by using questionably reliable data?

One clue might be found in Culicover and Jackendoff (2010), who write:

[T]heoreticians’ subjective judgments are essential in formulating linguistic theories. It would cripple linguistic investigation if it were required that all judgments of ambiguity and grammaticality be subject to statistically rigorous experiments on naive subjects. (Culicover and Jackendoff 2010)

The worry is that use of experimental methods is so resource consumptive that it would impede the formulation of linguistic theories. But this changes the subject from the importance of using reliable data as evidence in theory testing to using only experimentally gathered data in theory formulation . We are not aware of anyone who has ever suggested that at the stage of hypothesis development or theory formulation the linguist should eschew intuition. Certainly Bard et al., Schütze, Cowart, Gibson & Fedorenko, and Ferreira say no such thing. The relevant issue concerns what data should be used to test theories, which is a very different matter.

We noted earlier that there are clear and uncontroversial acceptability judgments, and that these judgments are reliable data. The difficulty lies in distinguishing the clear, uncontroversial, and reliable data from what only appears to be clear, uncontroversial, and reliable to a research community at a time. William Labov, the founder of modern quantitative sociolinguistics, who takes an Emergentist approach, proposed a set of working methodological principles in Labov (1975) for adjudicating when experimental methods should be employed.

The Consensus Principle : If there is no reason to think otherwise, assume that the judgments of any native speaker are characteristic of all speakers. The Experimenter Principle : If there is any disagreement on introspective judgments, the judgments of those who are familiar with the theoretical issues may not be counted as evidence. The Clear Case Principle : Disputed judgments should be shown to include at least one consistent pattern in the speech community or be abandoned. If differing judgments are said to represent different dialects, enough investigation of each dialect should be carried out to show that each judgment is a clear case in that dialect. (Labov 1975, quoted in Schütze 1996: 200)

If we accept that ‘introspective judgments’ are acceptability judgments, then Labov’s rules of thumb are guides for when to deploy experimental methods, although they no doubt need refinement. However, it seems vastly more likely that careful development of such methodological rules of thumb can serve to improve the reliability of linguistic data and adjudicate these methodological disputes that seem largely independent of any particular approach to linguistics.

In linguistics, the goal of collecting corpus data is to identify and organize a representative sample of a written and/or spoken variety from which characteristics of the entire variety or genre can be induced. Concordances of word usage in linguistic context have long been used to aid in the translation and interpretation of literary and sacred texts of particular authors (e.g. Plato, Aristotle, Aquinas) and of particular texts (e.g. the Torah, the rest of the Old Testament, the Gospels, the Epistles). Formal textual criticism, the identification of antecedently existing oral traditions that were later redacted into Biblical texts, and author identification (e.g. figuring out which of the Epistles were written by Paul and which were probably not) began to develop in the late 19th century.

The development of computational methods for collecting, analyzing, and searching corpora have seen rapid development as computer memory has become less expensive and search and analysis programs have become faster. The first computer searchable corpus of American English, the Brown Corpus, developed in the 1960s, contained just over one million word tokens. The British National Corpus (BNC) is a balanced corpus containing over 100 million words—a hundredfold size increase—of which 90% is written prose published from 1991 to 1994 and 10% is spoken English. Between 2005 and 2007, billion-word corpora were released for British English (ukWaC), German (deWaC), and Italian (itWaC)—a thousand times bigger than the Brown corpus. And the entire World Wide Web probably holds about a thousand times as much as that—around a trillion words. Thus corpus linguistics has gone from megabytes of data (∼ 10 3 kB) to terabytes of data (∼ 10 9 kB) in fifty years.

Just as a central issue concerning acceptability judgment data concerns its reliability as evidence for empirical generalizations about languages or idiolects, a central question concerning the collection of corpus data concerns whether or not it is representative of the language variety it purports to represent. Some linguists make the criterion of representativeness definitional: they call a collection of samples of language use a corpus only if it has been carefully balanced between different genres (conversation, informal writing, journalism, literature, etc.), regional varieties, or whatever.

But corpora are of many different kinds. Some are just very large compilations of text from individual sources such as newspapers of record or the World Wide Web—compilations large enough for the diversity in the source to act as a surrogate for representativeness. For example, a billion words of a newspaper, despite coming from a single source, will include not only journalists’ news reports and prepared editorials but also quoted speech, political rhetoric, humor columns, light features, theater and film reviews, readers’ letters, fiction items, and so on, and will thus provide examples of a much wider variety of styles than one might have thought.

Corpora are cleaned up through automatic or manual removal of such elements as numerical tables, typographical slips, spelling mistakes, markup tags, accidental repetitions ( the the ), larger-scale duplications (e.g., copies on mirror sites), boilerplate text ( Opinions expressed in this email do not necessarily reflect …), and so on (see Baroni et al. 2009 for a fuller discussion of corpus cleaning).

The entire web itself can be used as a corpus to some degree, despite its constantly changing content, its multilinguality, its many tables and images, and its total lack of quality control; but when it is, the outputs of searches are nearly always cleaned by disregarding unwanted results. For example, Google searches are blind to punctuation, capitalization, and sentence boundaries, so search results for to be will unfortunately include irrelevant cases, such as where a sentence like Do you want to? happens to be followed by a sentence like Be careful .

Corpora can be annotated in ways that permit certain kinds of analysis and grammar testing. One basic kind of annotation is part-of-speech tagging, in which each word is labeled with its syntactic category. Another is lemmatization, which classifies the different morphologically inflected forms of a word as belonging together ( goes , gone , going , and went belong with go , for example). A more thoroughgoing kind of annotation involves adding markup that encodes trees representing their structure; an example like That road leads to the freeway might be marked up as a Clause within which the first two words make up a Noun Phrase (NP), the last four constitute a Verb Phrase (VP), and so on, giving a structural analysis represented thus:

Structural analysis of 'That road leads to the highway'

Such a diagram is isomorphic to (and the one shown was computed directly from) a labeled bracketing like this:

(.Clause. (.NP. (.D. ‘that’ ) (.N. ‘road’ ) ) (.VP. (.V. ‘leads’ ) (.PP. (.P. ‘to’ ) (.NP. (.D. ‘the’ ) (.N. ‘freeway’ ) ) ) ) )

and this in turn could be represented in a markup language like XML as:

A corpus annotated with tree structure is known as a treebank . Clearly, such a corpus is not a raw record of attested utterances at all; it is a combination of a collection of attested utterances together with a systematic attempt at analysing their structure. Whether the analysis is added manually or semi-automatically, it is ultimately based on native speaker judgments. (Treebanks are often developed by graduate student annotators tutored by computational linguists; naturally, consistency between annotators is an issue that needs regular attention. See Artstein and Poesio, 2008, for discussion of the methodological issues.).

One of the purposes of a treebank is to permit the further investigation of a language and the checking of further linguistic hypotheses by searching a large database of previously established analyses. It can also be used to test grammars, natural language processing systems, or machine learning programs.

Going beyond syntactic parse trees, it is possible to annotate corpora further, with information of a semantic and pragmatic nature. There is ongoing computational linguistic research aimed at discovering whether, for example, semantic annotation that is semi-automatically added might suffice for recognition of whether a product review is positive or negative (what computational linguists call ‘sentiment analysis’).

Notice, then, that using corpus data does not mean abandoning or escaping from the use of intuitions about acceptability or grammatical structure: the results of a corpus search are generally filtered through the judgments of an investigator who decides which pieces of corpus data are to be taken at face value and which are just bad hits or irrelevant noise.

Difficult methodological issues arise in connection with the collection, annotation, and use of corpus data. For example, there is the issue of extremely rare expression tokens. Are they accurately recorded tokens of expression types that turn up only in consequence of sporadic errors and should be dismissed as irrelevant unless the topic of interest is performance errors? Are they due to errors in the compilation of the corpus itself, corresponding to neither accepted usage nor sporadic speech errors? Or are they perfectly grammatical but (for some extraneous reason) very rare, at least in that particular corpus?

Many questions arise about what kind of corpus is best suited to the research questions under consideration, as well as what kind of annotation is most appropriate. For example, as Ferreira (2005: 375) points out, some large corpora, insofar as they have not been cleaned of speech errors, provide relevant data for studying the distribution of speech disfluencies. In addition, probabilistic information about the relation between a particular verb and its arguments has been used to show that “verb-argument preferences [are] an essential part of the process of sentence interpretation” (Roland and Jurafsky 2002: 325): acceptability judgments on individual expressions do not provide information about the distribution of a verb and its arguments in various kinds of speech and writing. Studying conveyed meaning in context and identification of speech acts will require a kind of data that decontextualized acceptability judgments do not provide but semantically annotated corpora might.

Many Essentialists have been skeptical of the reliability of uncleaned, unanalyzed corpus data as evidence to support linguistic theorizing, because it is assumed to be replete with strings that any native speaker would judge unacceptable. And many Emergentists and Externalists, as well as some Essentialists, have charged that informally gathered acceptability judgments can be highly unreliable too. Both worries are apposite; but the former does not hold for adequately cleaned and analyzed corpora, and the latter does not hold for judgment data that has been gathered using appropriately controlled methods. In certain contested cases of acceptability, it will of course be important to use both corpus and controlled elicitation methods to cross-compare.

Notice that we have not in any way suggested that our three broad approaches to linguistics should differ in the kinds of data they use for theory testing: Essentialists are not limited to informal elicitation; nor are Emergentists and Externalists denied access to it. In matters of methodology, at least, there is in principle an open market—even if many linguists seem to think otherwise.

4. Language Acquisition

The three approaches to linguistic theorizing have at least something to say about how languages are acquired, or could in principle be acquired. Language acquisition has had a much higher profile since generative Essentialist work of the 1970s and 1980s gave it a central place on the agenda for linguistic theory.

Research into language acquisition falls squarely within the psychology of language; see the entry on language and innateness . In this section we do not aim to deal in detail with any of the voluminous literature on psychological or computational experiments bearing on language acquisition, or with any of the empirical study of language acquisition by developmental linguists, or the ‘stimulus poverty’ argument for the existence of innate knowledge about linguistic structure (Pullum and Scholz 2002). Our goals are merely to define the issue of linguistic nativism , set it in context, and draw morals for our three approaches from some of the mathematical work on inductive language learning.

The reader with prior acquaintance with the literature of linguistics will notice that we have not made reference to any partitioning of linguists into two camps called ‘empiricists’ and ‘rationalists’ (see e.g. Matthews 1984, Cowie 1999). We draw a different distinction relating to the psychological and biological prerequisites for first language acquisition. It divides nearly all Emergentists and Externalists from most Essentialists. It has often been confused with the classical empiricist/rationalist issue.

General nativists maintain that the prerequisites for language acquisition are just general cognitive abilities and resources. Linguistic nativists , by contrast, claim that human infants have access to at least some specifically linguistic information that is not learned from linguistic experience. Table 3 briefly sketches the differences between the two views.

Table 3: General and linguistic nativism contrasted

There does not really seem to be anyone who is a complete non-nativist: nobody really thinks that a creature with no unlearned capacities at all could acquire a language. That was the point of the much-quoted remark by Quine (1972: 95–96) about how “the behaviorist is knowingly and cheerfully up to his neck in innate mechanisms of learning-readiness”. Geoffrey Sampson (2001, 2005) is about as extreme an opponent of linguistic nativism as one can find, but even he would not take the failure of language acquisition in his cat to be unrelated to the cognitive and physical capabilities of cats.

The issue on which empirical research can and should be done is whether some of the unlearned prerequisites that humans enjoy have specifically linguistic content. For a philosophically-oriented discussion of the matter, see chapters 4–6 of Stainton (2006). For extensive debate about “the argument from poverty of the stimulus”, see Pullum and Scholz (2002) together with the six critiques published in the same issue of The Linguistic Review and the responses to those critiques by Scholz and Pullum (2002).

Linguists have given considerable attention to considerations of in-principle learnability —not so much the course of language acquisition as tracked empirically (the work of developmental psycholinguists) but the question of how languages of the human sort could possibly be learned by any kind of learner. The topic was placed squarely on the agenda by Chomsky (1965); and a hugely influential mathematical linguistics paper by Gold (1967)has dominated much of the subsequent discussion.

4.2.1 The Gold paradigm

Gold began by considering a reformulation of the standard philosophical problem of induction. The trouble with the question ‘Which hypothesis is correct given the totality of the data?’ is of course the one that Hume saw: if the domain is unbounded, no finite amount of data can answer the question. Any finite body of evidence will be consistent with arbitrarily many hypotheses that are not consistent with each other. But Gold proposed replacing the question with a very different one: Which tentative hypothesis is the one to pick , given the data provided so far, assuming a finite number of wrong guesses can be forgiven?

Gold assumed that the hypotheses, in the case of language learning, were generative grammars (or alternatively parsers; he proves results concerning both, but for brevity we follow most of the literature and neglect the very similar results on parsers). The learner’s task is conceived of as responding to an unending input data stream (ultimately complete, in that every expression eventually turns up) by enunciating a sequence of guesses at grammars.

Although Gold talks in developmental psycholinguistic terms about language learners learning grammars by trial and error, his extremely abstract proofs actually make no reference to the linguistic content of languages or grammars at all. The set of all finite grammars formulable in any given metalanguage is computably enumerable, so grammars can be systematically numbered. Inputs—grammatical expressions from the target language—can also be numerically encoded. We end up being concerned simply with the existence or non-existence of certain functions from natural number sequences to natural numbers.

A successful learner is one who uses a procedure that is guaranteed to eventually hit on a correct grammar. For single languages, this is trivial: if the target language is L and it is generated by a grammar G , then the procedure “Always guess G ” does the job, and every language is learnable. What makes the problem interesting is applying it to classes of grammars. A successful learner for a class C is one who uses a procedure that is guaranteed to succeed no matter what grammar from C is the target and no matter what the data stream is like (as long as it is complete and contains no ungrammatical examples).

Gold’s work has interesting similarities with earlier philosophical work on inductive learning by Hilary Putnam (1963; it is not clear whether Gold was aware of this paper). Putnam gave an informal proof of a sort of incompleteness theorem for inductive regularity-learning devices: no matter what algorithm is used in a machine for inducing regularities from experience, and thus becoming able to predict events, there will always be some possible environmental regularities that will defeat it. (As a simple example, imagine an environment giving an unbroken sequence of presentations all having some property a . If there is a positive integer n such that after n presentations the machine will predict that presentation number n + 1 will also have property a , then the machine will be defeated by an environment consisting of n presentations of a followed by one with the incompatible property b —the future need not always resemble the past. But if on the other hand there is no such n , then an environment consisting of an unending sequence of a presentations will defeat it.)

Gold’s theorems are founded on certain specific idealizing assumptions about the language learning situation, some of which are intuitively very generous to the learner. The main ones are these:

  • Pre-set grammar class . A class of grammars from among which to select is fixed ab initio, and the learner’s strategy can be one that only works for that class.
  • Pre-set vocabulary . A finite universal vocabulary of elements V is fixed ab initio, and the learner can rely on not encountering any other elements (though the learner does not know which subset of V is used in the target language).
  • Unending input . The input (the evidence presented to the learner) goes on forever—though it may contain arbitrary repetitions, and a successful learner will always reach a point where no future evidence will cause a change of guess.
  • Exhaustive evidence . Ultimately every expression in the language will appear in the evidence presented to the learner.
  • No noise . Every input example is a grammatical expression of the target language.
  • No ordering restrictions . Any expression may appear at any point in the input data stream.
  • No memory limit . The learner can remember every expression ever presented.
  • No time limit . Learning must be achieved after some finite time, but no fixed bound is set in advance.
  • Generative grammar target . What is ultimately learned is a generative grammar.
  • No statistics . Frequency of particular expressions in the input plays no role in the learning process.

The most celebrated of the theorems Gold proved (using some reasoning remarkably similar to that of Putnam 1963) showed that a language learner could be similarly hostage to malign environments. Imagine a learner being exposed to an endless and ultimately exhaustive sequence of presented expressions from some target language—Gold calls such a sequence a ‘text’. Suppose the learner does not know in advance whether the language is infinite, or is one of the infinitely many finite languages over the vocabulary V . Gold reasons roughly thus:

  • There must be some n such that an environment consisting of a sequence of n presented expressions all taken from a certain finite language L 1 (possibly with many repetitions) will cause the learner to guess the target language is L 1 . (If there is not, then we already know how to baffle the learner: the learner will be unable to learn L 1 from any text.)
  • But if there is such an n , then the learner will be baffled by any infinite target language that is a superset of them all: a text consisting of n presentations of expressions from L 1 followed by n presentations of a slightly larger finite language L 2 , and so on forever (there is no largest finite language, and ex hypothesi the learner will keep trying them all).

Leaping too soon to the conclusion that the target language is infinite will be disastrous, because there will be no way to retrench: no presented examples from a finite language L k will ever conflict with the hypothesis that the target is some infinite superset of L k .

The relevance of all this to the philosophy of linguistics is that the theorem just sketched has been interpreted by many linguists, psycholinguists, and philosophers as showing that humans could not learn languages by inductive inference based on examples of language use, because all of the well-known families of languages defined by different types of generative grammar have the crucial property of allowing grammars for every finite language and for at least some infinite supersets of them. But Gold’s paper has often been over-interpreted. A few examples of the resultant mistakes follow.

It’s not about underdetermination . Gold’s negative results are sometimes wrongly taken to be an unsurprising reflection of the underdetermination of theories by finite bodies of evidence (Hauser et al. 2002 seem to make this erroneous equation on p. 1577; so do Fodor and Crowther 2002, implicitly—see Scholz and Pullum 2002, 204–206). But the failure of text-identifiability for certain classes of languages is different from underdetermination in a very important way, because there are infinite classes of infinite languages that are identifiable from text. The first chapter of Jain et al. (1999) discusses an illustrative example (basically, it is the class containing, for all n > 0, the set of all strings with length greater than n ). There are infinitely many others. For example, Shinohara (1990) showed that for any positive integer n the class of all languages generated by a context-sensitive grammar with not more than n rules is learnable from text.

It’s not about stimulus poverty . It has also sometimes been assumed that Gold is giving some kind of argument from poverty of the stimulus (there are signs of this in Cowie 1999, 194ff; Hauser et al. 2002, 1577; and Prinz 2002, 210). This is very clearly a mistake (as both Laurence and Margolis 2001 and Matthews 2007 note): in Gold’s text-learning scenario there is no stimulus poverty at all. Every expression in the language eventually turns up in the learner’s input.

It’s not all bad news . It is sometimes forgotten that Gold established a number of optimistic results as well as the pessimistic one about learning from text. Given what he called an ‘informant’ environment rather than a text environment, we see strikingly different results. An informant environment is an infinite sequence of presentations sorted into two lists, positive instances (expressions belonging to the target language) and negative instances (not in the language). Almost all major language-theoretic classes are identifiable in the limit from an informant environment (up to and including the class of all languages with a primitive recursive characteristic function, which comes close to covering any language that could conceivably be of linguistic interest), and all computably enumerable languages become learnable if texts are allowed to be sequenced in particular ways (see the results in Gold 1967 on ‘anomalous text’).

Gold did not give a necessary condition for a class to be identifiable in the limit from text, but Angluin (1980) later provided one (in a result almost but not quite obtained by Wexler and Hamburger 1973). Angluin showed that a class C is text-identifiable iff every language L in C has a finite “telltale” subset T such that if T is also proper subset of some other language in C , that other language is not a proper subset of L . This condition precludes guessing too large a language. Once all the members of the telltale subset for L have been received as input, the learner can safely make L the current conjecture. The language to be identified must be either L or (if subsequent inputs include new sentences not in L ) some larger language, but it can’t be a proper subset of L .

Johnson (2004) provides a useful review of several other misconceptions about Gold’s work; e.g., the notion that it might be the absence of semantics from the input that makes identification from text impossible (this is not the case).

4.2.2 Gold’s paradox

Some generative Essentialists see a kind of paradox in Gold’s results—a reductio on one or more of the assumptions he makes about in-principle learnability. To put it very crudely, learning generative grammars from presented grammatical examples seems to have been proved impossible, yet children do learn their first languages, which for generative Essentialists means they internalize generative psychogrammars, and it is claimed to be an empirical fact that they get almost no explicit evidence about what is not in the language (Brown and Hanlon 1970 is invariably cited to support this). Contradiction. Gold himself suggested three escape routes from the apparent paradox:

  • Assume tighter limits on the pre-set grammar class. Perhaps, for example, learners have an ‘innate’ grasp of some definition of the pre-set grammar class that guarantees its learnability. (For example, identifiability in the limit from text could be guaranteed by ensuring that the class of candidate languages does not contain both (a) some infinite set of finite languages and (b) some infinite language that is the union of all of them.)
  • Assume learners get systematic information about what is not in the language (perhaps indirectly, in ways not yet recognized), so that the environment is of the informant type rather than the text type.
  • Assume some helpful feature is present in learning environments. The ‘no order restrictions’ assumption might be false: there could be regularities in the order of expressions in texts that can support inferences about what is ungrammatical.

All three of these paths have been subsequently explored. Path (1) appealed to generative Essentialists. Chomsky (1981) suggested an extreme restriction: that universal grammar permitted only finitely many grammars. This claim (for which Chomsky had little basis: see Pullum 1983) would immediately guarantee that not all finite languages are humanly learnable (there are infinitely many finite languages, so for most of them there would be no permissible grammar). Osherson and Weinstein (1984) even proved that under three fairly plausible assumptions about the conditions on learning, finiteness of the class of languages is necessary—that is, a class must be finite if it is to be identifiable from text. However, they also proved that this is not sufficient: there are very small finite classes of languages that are not identifiable from text, so it is logically possible for text-identification to be impossible even given only a finite number of languages (grammars). These two results show that Chomsky’s approach cannot be the whole answer.

Path (2) proposes investigation of children’s input with an eye to finding covert sources of negative evidence. Various psycholinguists have pursued this idea; see the entry on language and innateness in this encyclopedia, and (to cite one example) the results of Chouinard and Clark (2003) on hitherto unnoticed sources of negative evidence in the infant’s linguistic environment, such as parental corrections.

Path (3) suggests investigating the nature of children’s linguistic environments more generally. Making evidence available to the learner in some fixed order can certainly alter the picture quite radically (Gold proved that if some primitive-recursive generator controls the text it can in effect encode the identity of the target language so that all computably enumerable languages become identifiable from text). It is possible in principle that limitations on texts (or on learners’ uptake) might have positive rather than negative effects on learnability (see Newport 1988; Elman 1993; Rohde and Plaut 1999; and the entry on language and innateness ).

4.2.3 The claimed link to ‘rationalism’ versus ‘empiricism’

Gold’s suggested strategy of restricting the pre-set class of grammars has been interpreted by some as a defense of rationalist rather than empiricist theories of language acquisition. For example, Wexler and Culicover state:

Empiricist theory allows for a class of sensory or peripheral processing mechanisms by means of which the organism receives data. In addition, the organism possesses some set of inductive principles or learning mechanisms…Rationalist theory also assumes that a learner has sensory mechanisms and inductive principles. But rationalist theory assumes that in addition the learner possesses a rich set of principles concerning the general nature of the ability that is to be learned. (Wexler and Culicover 1980: 5)

Wexler and Culicover claim that ‘empiricist’ learning mechanisms are both weak and general: not only are they ‘not related to the learning of any particular subject matter or cognitive ability’ but they are not ‘limited to any particular species’. It is of course not surprising that empiricist learning fails if it is defined in a way that precludes drawing a distinction between the cognitive abilities of humans and fruit flies.

Equating Gold’s idea of restricting the class of grammars with the idea of a ‘rationalist’ knowledge acquisition theory, Wexler and Culicover try to draw out the consequences of Gold’s paradigm for the Essentialist linguistic theory of Chomsky (1965). They show how a very tightly restricted class of transformational grammars could be regarded as text-identifiable under extremely strong assumptions (e.g., that all languages have the same innately known deep structures).

Matthews (1984) follows Wexler and Culicover’s lead and draws a more philosophically oriented moral:

The significance of Gold’s result becomes apparent if one considers that (i) empiricists assume that there are no constraints on the class of possible languages (besides perhaps that natural languages be recursively enumerable), and (ii) the learner employs a maximally powerful learning strategy—there are no strategies that could accomplish what that assumed by Gold cannot. These two facts, given Gold’s unsolvability result for text data, effectively dispose of the empiricist claim that there exists a ‘discovery procedure’. (1989: 60)

The actual relation of Gold’s results to the empiricism/rationalism controversy seems to us rather different. Gold’s paradigm looks a lot more like a formalization of so-called ‘rationalism’. The fixed class of candidate hypotheses (grammars) corresponds to what is given by universal grammar—the innate definition of the essential properties of language. What Gold actually shows, therefore, is not “the plausibility of rationalism” but rather the inadequacy of a huge range of rationalist theories: under a wide range of different choices of universal grammar, language acquisition appears to remain impossible.

Moreover, Matthews ignores (as most linguists have) the existence of large and interesting classes of languages that are text-identifiable.

Gold’s result, like Putnam’s earlier one, does show that a certain kind of trial-and-error inductive learning is insufficient to permit learning of arbitrary environmental regularities. There has to be some kind of initial bias in the learning procedure or in the data. But ‘empiricism’, the supposed opponent of ‘rationalism’, is not to be equated with a denial of the existence of learning biases. No one doubts that humans have inductive biases. To quote Quine again, “Innate biases and dispositions are the cornerstone of behaviorism, and have been studied by behaviorists” (1972: 95–96). As Lappin and Shieber (2007) stress, there cannot be such a thing as a learning procedure (or processing mechanism) with no biases at all.

The biases posited in Emergentist theories of language acquisition are found, at least in part, in the non-linguistic social and cognitive bases of human communication. And the biases of Externalist approaches to language acquisition are to be found in the distributional and stochastic structure of the learning input and the multitude of mechanisms that process that input and their interactions. All contemporary approaches to language acquistion have acknowledged Gold’s results, but those results do not by themselves vindicate any one of our three approaches to the study of language.

Gold’s explicit equation of acquiring a language with identifying a generative grammar that exactly generates it naturally makes his work seem relevant to generative Essentialists (though even for them, his results do not provide anything like a sufficient reason for adopting the linguistic nativist position). But another key assumption, that nothing about the statistical structure of the input plays a role in the acquisition process, is being questioned by increasing numbers of Externalists, many of whom have used Bayesian modeling to show that the absence of positive evidence can function as a powerful source of (indirect) negative evidence: learning can be driven by what is not found as well as by what is (see e.g. Foraker et al. (2009)).

Most Emergentists simply reject the assumption that what is learned is a generative grammar. They see the acquisition task as a matter of learning the details of an array of constructions (roughly, meaning-bearing ways of structurally composing words or phrases) and how to use them to communicate. How such learning is accomplished needs a great deal of further study, but Gold’s paper did not show it to be impossible.

5. Language Evolution

Over the past three decades a large amount of work has been done on topics to which the term ‘language evolution’ is attached, but there are in fact four distinct such topics:

  • the phylogenetic emergence of non-human communication capacities, systems, and behaviors in various animals;
  • the phylogenetic emergence of uniquely human communication capacities, systems, and behaviors;
  • the phylogenetic emergence, unique in humans, of the capacity (or capacities) to develop , acquire , and use language;
  • the course of historical evolution through intergenerational changes in particular languages as they are acquired and used by humans.

Emergentists tend to regard any of the topics (a)–(d) as potentially relevant to the study of language evolution. Essentialists tend to focus solely on (c). Some Essentialists even deny that (a) and (b) have any relevance to the study of (c); for example:

There is nothing useful to be said about behavior or thought at the level of abstraction at which animal and human communication fall together… [H]uman language, it appears, is based on entirely different principles. This, I think, is an important point, often overlooked by those who approach language as a natural, biological phenomenon; in particular, it seems rather pointless, for these reasons, to speculate about the evolution of human language from simpler systems… (Chomsky 1968: 62)

Other generative Essentialists, like Pinker and Bloom (1990) and Pinker and Jackendoff (2005), seem open to the view that even the most elemental aspects of topic (b) can be directly relevant to the study of (c). This division among Essentialists reflects a division among their views about the role of adaptive explanations in the emergence of (b) and especially (c). For example:

We know very little about what happens when 10 10 neurons are crammed into something the size of a basketball, with further conditions imposed by the specific manner in which this system developed over time. It would be a serious error to suppose that all properties, or the interesting properties of the structures that evolved, can be ‘explained’ in terms of ‘natural selection’. (Chomsky 1975:59, quoted by Newmeyer 1998 and Jackendoff 2002)

The view expressed here that all (or even most) interesting properties of the language faculty are not adaptations conflicts with the basic explanatory strategy of evolutionary psychology found in the neo-Darwinian Essentialist views of Pinker and Bloom. Piattelli-Palmarini (1989), following Chomsky, adopts a fairly standard Bauplan critique of adaptationism. On this view the language faculty did not originate as an adaptation, but more plausibly “may have originally arisen for some purely architectural or structural reason (perhaps overall brain size, or the sheer duplication of pre-existing modules), or as a by product of the evolutionary pressures” (p. 19), i.e., it is a kind of Gouldian spandrel.

More recently, some Essentialist-leaning authors have rejected the view that no analogies and homologies between animal and human communication are relevant to the study of language. For example, in the context of commenting on Hauser et al. (2002), Tecumseh Fitch (2010) claims that “Although Language, writ large, is unique to our species (many probably most) of the mechanisms involved in language have analogues or homologues in other animals.” However, the view that the investigation of animal communication can shed light on human language is still firmly rejected by some. For example, Bickerton (2007: 512) asserts that “nothing resembling human language could have developed from prior animal call systems.”

Bickerton fronts the following simple argument for his view:

If any adaptation is unique to a species, the selective pressure that drove it must also be unique to that species; otherwise the adaptation would have appeared elsewhere, at least in rudimentary form. (2007: 514)

Thus, the mere fact that language is unique to humans is sufficient to rule out monkey and primate call systems as preadapations for language. But, contra Bickerton, a neo-Darwinian like Jackendoff (2002) appeals to the work of Dunbar (1998), Power (1998), Worden (1998) to provide a selectionist story which assumes that cooperation in hunting, defense (Pinker and Bloom 1990), and “ ‘social grooming’ or deception” are selective forces that operated on human ancestors to drive increases in expressive power that distinguishes non-human communication and human linguistic capacities and systems. Bickerton (2014), however, combines aspects of Essentialism, Emergentism, and Externalism by taking equal parts of Minimalism, primatology, and cultural evolution into a more holistic account. He specifically tailors a niche construction theory to explain the emergence of displaced, discrete symbolization in a particular kind of primate, namely human beings. He thus allows for (a) and (b) to figure in an explanation of (c). This is somewhat of a departure from his earlier positions.

Within the general Essentialist camp, language evolution has taken center stage since the inception of the Minimalist Program. An explanation of the evolution of language now became one of the main theoretical driving forces behind linguistic theory and explanation. Again, the focus seems to have stayed largely on (c). Berwick and Chomsky explicitly state:

At some time in the very recent past, apparently sometime before 80,000 years ago, if we can judge from associated symbolic proxies, individuals in a small group of hominids in East Africa underwent a minor biological change that provided the operation Merge-an operation that takes human concepts as computational atoms and yields structured expressions that, systematically interpreted by the conceptual system, provide a rich language of thought. (2016: 87)

Such theories rely heavily on the possibility of the evolution of language being explained in terms of saltation or random mutation. This postulate has come under significant scrutiny (see Steedman 2017). Saltation views, however, rely on one of the core assumptions mentioned in the quote above, i.e. that language evolved circa 100 000 years old. This central claim has recently been challenged by Everett (2017) who cites paleontological evidence from the alleged nautical abilities of Homo Erectus to dismantle this timeline. If true, this would mean that language evolved around two million years ago and random mutation need not be the only viable explanation as many in the Essentialist framework assume (see Progovac 2015 for a particular gradualist account).

While generative Essentialists debate among themselves about the plausibility of adaptive explanations for the emergence of essential features of a modular language capacity, Emergentists are perhaps best characterized as seeking broad evolutionary explanations of the features of languages (topic (c)) and communicative capacities (topics (b) and (c)) conceived in non-essentialist, non-modular ways. And as theorists who are committed to exploring non-modular views of linguistic capacities (topic (c)), the differences and similarities between (a) and (b) are potentially relevant to (c).

Primatologists like Cheney and Seyfarth, psychologists like Tomasello, anthropologists like Terrence Deacon, and linguists like Phillip Lieberman share an interest in investigating the communicative, anatomical, and cognitive characteristics of non-human animals to identify biological differences between humans, and monkeys and primates. In the following paragraph we discuss Cheney and Seyfarth (2005) as an example, but we could easily have chosen any of a number of other theorists.

Cheney and Seyfarth (2005) emphasize that non-human primates have a small, stimulus specific repertoire of vocal productions that are not “entirely involuntary,” and this contrasts with their “almost openended ability to learn novel sound-meaning pairs” (p. 149). They also emphasize that vocalizations in monkeys and apes are used to communicate information about the vocalizer, not to provide information intended to “rectify false beliefs in others or instruct others” (p. 150). Non-human primate communication consists in the mainly involuntary broadcasting of the vocalizer’s current affective state. Moreover, although Cheney and Seyfarth recognize that the vervet monkey’s celebrated call system (Cheney and Seyfarth 1990) is “functionally referential” in context, their calls have no explicit meaning since they lack “any propositional structure”. From this they conclude:

The communication of non-human animals lacks three features that are abundantly present in the utterances of young children: a rudimentary ability to attribute mental states different from their own to others, the ability to generate new words, and lexical syntax. (2005: 151)

By ‘lexical syntax’ Cheney and Seyfarth mean a kind of semantic compositionality of characteristic vocalizations. If a vocalization (call) were to have lexical syntax, the semantic significance of the whole would depend on the relation of the structure of parts of the call to the structure of what they signify. The absence of ‘lexical syntax’ in call systems suggests that it is illegitimate to think of them as having anything like semantic structure at all.

Despite the rudimentary character of animal communication systems when compared with human languages, Cheney and Seyfarth argue that monkeys and apes exhibit at least five characteristics that are pre-adaptations for human communication:

  • their vocalizations are representational;
  • they have competitive/cooperative relations in which alliances, friendships, and rivalries that “create selective pressures for the kind of complex, abstract conceptual abilities that are likely to have proceeded the earlier linguistic communication”;
  • because of (ii), their representations of social relations between individuals and themselves are hierarchally structured;
  • certain monkeys, e.g. baboons, have open-ended, rule-governed systems of social knowledge;
  • their knowledge is propositional.

It is, of course, controversial to claim that monkeys have rule-governed propositional social knowledge systems as claimed in (iv) and (v). For instance, Tomasello’s (2008) ‘Cooperative Communication’ approach makes a case for primate intentional systems not based on their vocalizations but on their gestural systems. Therein he claims that “great ape gestural communication shares with human linguistic communication foundational aspects of its manner of functioning, namely, the intentional and flexible use of learned communicative signals” (2008: 21).

But Emergentists, Externalists, and Essentialists could all, in principle, agree that there are both unique characteristics of human communicative capacities and characteristics of such capacities that are shared with non-humans. For example, by the age of one, human infants can use direction of gaze and focus of attention to infer the referent of a speaker’s utterance (Baldwin and Moses 1994). By contrast, this sort of social referencing capacity in monkeys and apes is rudimentary. This suggests that a major component of humans’ capacity to infer a specific referent is lacking in non-humans.

Disagreements between the approaches might be due to the perceived significance of non-human communicative capacities and their relation to uniquely human ones.

We mentioned earlier that both early 20th-century linguistics monographs and contemporary introductory textbooks include discussions of historical linguistics, i.e., that branch that studies the history and prehistory of changes in particular languages, how they are related to each other, and how and why they change. Again, this topic is distinct from the emergence of language in hominoid species and concerns mostly the linguistic changes that have occurred over a much shorter period within languages.

The last decade has seen two kinds of innovations related to studying changes in particular languages. One, which we will call ‘linguistic phylogeny’, concerns the application of stochastic phylogenetic methods to investigate prehistoric population and language dispersion (Gray and Jordan 2000, Gray 2005, Atkinson and Gray 2006, Gray et al. 2009). These methods answer questions about how members of a family of languages are related to each other and dispersed throughout a geographic area. The second, which we will call the effects of transmission, examines how interpreted artificial languages (sets of signifier/signified pairs) change under a range of transmission conditions (Kirby et al. 2008, Kirby 2001, Hurford 2000), thus providing evidence about how the process of transmission affects the characteristics, especially the structure, of the transmitted interpreted system.

5.2.1 Linguistic phylogeny

Russell Gray and his colleagues have taken powerful phylogenetic methods that were developed by biologists to investigate molecular evolution, and applied them to linguistic data in order to answer questions about the evolution of language families. For example, Gray and Jordan (2000) used a parsimony analysis of a large language data set to adjudicate between competing hypotheses about the speed of the spread of Austronesian languages through the Pacific. More recently, Greenhill et al. (2010) used a NeighbourNet analysis to evaluate the relative rates of change in the typological and lexical features of Austronesian and Indo-European. These results bear on hypotheses about the relative stability of language types over lexical features of those languages, and how far back in time that stability extends. If there were highly conserved typological and lexical features, then it might be possible to identify relationships between languages that date beyond the 8000 (plus or minus 2000) year limit that is imposed by lexical instability.

5.2.2 Effects of transmission

The computational and laboratory experiments of Kirby and his collaborators have shown that under certain conditions of iterated learning, any given set of signifier/signified pairs in which the mapping is initially arbitrary will change to exhibit a very general kind of compositional structure. Iterated learning has been studied in both computational and laboratory experiments by means of diffusion chains, i.e., sequences of learners. A primary characteristic of such sequences of transmission is that what is transmitted from learner to learner will change in an iterated learning environment, in a way that depends on the conditions of transmission.

The children’s game called ‘Telephone’ in the USA (‘Chinese Whispers’ in the UK), provides an example of diffusion chains under which what is transmitted is not stable. In a diffusion chain learning situation what a chain member has actually learned from an earlier member of the chain is presented as the input to the next learner, and what that learner has actually learned provides the input to the following learner. In cases where the initial learning task is very simple: i.e., where what is transmitted is both simple, completely transmitted, and the transmission channel is not noisy, what is transmitted is stable over iterated transmissions even in cases when the participants are young children and chimpanzees (Horner et al. 2006). That is, there is little change in what is transmitted over iterated transmissions. However, in cases where what is transmitted is only partially presented, very complex, or the channel is noisy, then there is a decrease in the fidelity of what is transmitted across iterations just like there is in the children’s game of Telephone.

What Kirby and colleagues show is that when the initial input to a diffusion chain is a reasonably complex set of arbitrary signal/signifier pairs, e.g. one in which 27 complex signals of 6 letters are randomly assigned to 27 objects varying on dimensions of color, kind of motion, and shape, what is transmitted becomes more and more compositional over iterated transmission. Here, ‘compositional’ is being used to refer to the high degree to which sub-strings of the signals come to be systematically paired with specific phenomenal sub-features of what is signified. The transmission conditions in these experiments were free of noise, and for each iteration of the learning task only half of the possible 27 signifier/signified pairs were presented to participants. Under this kind of transmission bottleneck a high degree of sign/signified structure emerged.

A plausible interpretation of these results is that the developing structure of the collection of signs is a consequence of the repeated forced inference by participants from 14 signs and signifieds in the training set to the entire set of 27 pairs. A moral could be that iterated forced prediction of the sign/signified pairs in the entire set, on the basis of exposure to only about half of them, induced the development of a systematic, compositional structure over the course of transmission. It is reasonable to conjecture that this resulting structure reflects effects of human memory, not a domain-specific language module—although further work would be required to rule out many other competing hypotheses.

Thus Kirby and his colleagues focus on something very different from the prerequisites for language emergence. Linguistic nativists have been interested in how primates like us could have become capable of acquiring systems with the structural properties of natural languages. Kirby and his colleagues (while not denying that human cognitive evolution is of interest) are studying how languages evolve to be capable of being acquired by primates like us .

5.2.3 Trends in the Philosophy of Language Evolution

Lastly, language evolution has amassed a great deal of interdisciplinary work in recent times. This has allowed philosophers to directly contribute to this emerging field. The trends in the philosophical work have only loosely followed the Externalist, Emergentist and Essentialist divisions we advocate here. Most philosophical work has largely been focused on Emergentist conceptions within the evolution of linguistic meaning specifically.

Bar-On (2013) distinguishes between Gricean and Post-Gricean approaches to the evolution of language. The former requires an attribution of Gricean speaker meaning to our languageless ancestors which in turn seems to assume intentional actions govern by rationality (‘nonnatural meaning’). This task is as fraught as explaining the evolution of language itself. She thus proposes the latter, specifically a Post-Gricean (Orrigi and Sperber 2000) approach which takes expressive communication (found widely in non-human animal species) as a basis for the emergence of linguistic meaning between signalers and receivers. She states:

Expressive communication, I will argue, exhibits features that foreshadow significant aspects of linguistic communication. In its domain, we can identify legitimate natural precursors of meaningful linguistic communication. (For present purposes, by ‘legitimate natural precursors’, I mean behavioral interactions that at least: a. can be found in the natural world; b. go beyond Tomasello’s mere ‘communicative displays’; c. do not depend on crediting the relevant creatures with language-like propositional thought or post-Gricean communicative intentions, and; d. exhibit features that foreshadow important semantic and pragmatic features of linguistic communication so in that sense are proto-semantic and proto-pragmatic.) (2013: 354)

Recent work in Evolutionary Game Theory has also lent credence to the emergence of signaling systems involving non-intentional states. Taking Lewis (1969) as a spring-board, Skyrms (2010) investigates the structure of signaling behavior beyond the existence of mutual conventions. His framework starts from the most basic non-trivial cases and gradually introduces complexity (such as deception and the introduction of new signals etc.). Skyrms’ account views propositional or semantic content as a special case of informational content thereby reintroducing information theory as a tool to philosophers of language and linguistics interested in the emergence of linguistic communication and/or semantic meaning.

  • Adger, David, 2003, Core Syntax: A Minimalist Approach , New York: Oxford University Press.
  • Akmajian, Adrian, Demers, Richard, Farmer, Ann, and Harnish, Robert, 2010, Linguistics: An Introduction to Language and Communication , Cambridge, Massachusetts: MIT Press, 6th ed.
  • Angluin, Dana, 1980, “Inductive inference of formal languages from positive data”, Information and Control , 45: 117–135.
  • Artstein, Ron and Poesio, Massimo, 2008, “Inter-Coder Agreement for Computational Linguistics”, Computational Linguistics , 34: 555–596.
  • Atkinson, Quentin D. and Gray, Russell D., 2006, “How old is the Indo-European language family? Progress or more moths to the flame?”, in J. Clackson, P. Forster, and C. Renfrew, (eds.), Phylogenetic Methods and the Prehistory of Languages , Cambridge: MacDonald Institute for Archaeological Research, 91–109.
  • Azzouni, Jody, 2004, Deflating Existential Consequence: A Case for Nominalism , New York: Oxford University Press.
  • Baldwin, Dare A. and Moses, L. J., 1994, “Early understanding of referential intent and attentional focus: Evidence from language and emotion”, in C. Lewis and P. Mitchell (eds.), Children’s Early Understanding of Mind: Origins and Development , Hillsdale, NJ: Lawrence Erlbaum, 133–156.
  • Bard, Ellen, Robertson, David, and Sorace, Antonella, 1996, “Magnitude estimation of linguistic acceptability”, Language , 72(1): 32–68.
  • Barkow, J. H., Cosmides, Leda, and Tooby, J., 1992, The Adapted Mind: Evolutionary Psychology and the Generation of Culture , New York: Oxford University Press.
  • Barlow, Michael and Kemmer, Suzanne (eds.), 2002, Usage-Based Models of Language , Stanford: CSLI Press.
  • Bar-On, Dorit, 2013, “Origins of Meaning: Must We ‘Go Gricean’?” Mind and Language , 38(3): 342–375.
  • Baroni, M., Bernardini, S., Ferraresi, A., and Zanchetta, E., 2009, “The WaCky Wide Web: A collection of very large linguistically processed web-crawled corpora”, Journal of Language Resources and Evaluation , 43(3): 209–226.
  • Bates, Elizabeth, Elman, Jeffrey, Johnson, Mark, Karmiloff-Smith, Annette, Parisi, Domenico, and Plunkett, Kim, 1998, “Innateness and emergentism”, William Bechtel and George Graham (eds.), A Companion to Cognitive Science , Oxford: Basil Blackwell, 590–601.
  • Berwick, Robert, and Chomsky, Noam, 2016, Why only us? Language and evolution , Cambridge, MA: MIT Press.
  • Bickerton, Derek, 2007, “Language evolution: A brief guide for linguists”, Lingua , 117: 510–526.
  • Bickerton, Derek, 2014, More than nature needs: Language, mind, and evolution , Cambridge, MA: Harvard University Press.
  • Bloomfield, Leonard, 1914, An Introduction to the Study of Language , New York, NY: Henry Holt.
  • –––, 1933, Language , New York, NY: Henry Holt.
  • –––, 1939, Linguistic Aspects of Science (International Encyclopedia of Unified Science: Volume 1/Number 4), Chicago: University of Chicago Press.
  • Boroditsky, Lera, Schmidt, L., and Phillips, W., 2003, “Sex, syntax, and semantics”, in Dedre Gentner and Susan Goldin-Meadow (eds.), Language in Mind: Advances in the Study of Language and Cognition , Cambridge, Massachusetts: MIT Press, 61–80.
  • Bresnan, Joan, 2007, “A few lessons from typology”, Linguistic Typology , 11: 297–306.
  • Bresnan, Joan, Cueni, Anna, Nikitina, Tatiana, and Baayen, Harald, 2007, “Predicting the dative alternation”, in G. Boume, I. Kraemer, and J. Zwarts (eds.), Cognitive Foundations of Interpretation , Amsterdam: Royal Netherlands Academy of Science, 69–94.
  • Bresnan, Joan and Ford, Marilyn, 2010, “Predicting syntax: Processing dative constructions in American and Australian varieties of English”, Language , 86: 168–213.
  • Bromberger, Sylvain, 1989, “Types and tokens in linguistics”, in Reflections on Chomsky , A. George, (ed.), Oxford: Blackwell, 58–88.
  • –––, 2011, “What are words? Comments on Kaplan (1990), on Hawthorne and Lepore, and on the issue”, The Journal of Philosophy , 108(9): 486–503.
  • Brown, Roger and Hanlon, Camille, 1970, “Derivational complexity and order of acquisition in child speech”, in Cognition and the Development of Language , J. R. Hayes, (ed.), New York: John Wiley and Sons, 11–54.
  • Brown, Roger and Lenneberg, Eric, 1954, “A study in language and cognition”, Journal of Abnormal and Social Psychology , 49: 445–453.
  • Bybee, Joan and McClelland, J. L., 2005, “Alternatives to the combinatorial paradigm of linguistic theory based on domain general principles of human cognition”, The Linguistic Review , 22(2–4): 381–410.
  • Cappelen, Herman, 1999, “Intentions in words”, Noûs , 33(1): 92–102.
  • Cheney, D. L. and Seyfarth, R. M., 1990, “The assessment by vervet monkeys of their own and another species’ alarm calls”, Animal Behavior , 40: 754–764.
  • –––, 2005, “Constraints and preadaptations in the earliest stages of language evolution”, The Linguistic Review , 22: 135–159.
  • Chierchia, Gennaro, 1998, “Reference to kinds across languages”, Natural Language Semantics , 6: 339–405.
  • Chomsky, Noam, 1955, “The logical structure of linguistic theory”, Unpublished manuscript; revised in 1956 and distributed from MIT Library; published with some abridgement in 1975 by Plenum Press, New York.
  • –––, 1959, “On certain formal properties of grammars”, Information and Control , 1: 91–112.
  • –––, 1965, Aspects of the Theory of Syntax , Cambridge, Massachusetts: MIT Press.
  • –––, 1968, Language and Mind , New York: Harper and Row.
  • –––, 1969, “Linguistics and philosophy”, in Language and Philosophy: A Symposium, Sidney Hook, (ed.), New York: New York University Press, 51–94.
  • –––, 1975, Reflections on Language , New York, NY: Pantheon.
  • –––, 1979, Language and Responsibility , [translated by John Viertel from the 1977 French edition produced by Mitsou Ronat], Hassocks, Sussex: Harvester Press.
  • –––, 1981, Lectures on Government and Binding , Dordrecht: Foris.
  • –––, 1986, Knowledge of Language: Its Nature, Origin and Use , Westport, CT: Praeger.
  • –––, 1988, Language and Problems of Knowledge , Cambridge, Massachusetts: MIT Press.
  • –––, 1992, “Explaining language use”, Philosophical Topics , 20: 205–231.
  • –––, 1995, “Language and nature”, Mind , 104: 1–61.
  • –––, 2003, “Reply to Millikan”, in Louise M. Antony and Norbert Hornstein (eds.), Chomsky and His Critics , Oxford: Blackwell, 308–315.
  • Chouinard, M. M. and Clark, E. V., 2003, “Adult reformulations of child errors as negative evidence”, Journal of Child Language , 30(3): 637–669.
  • Cowart, Wayne, 1997, Experimental Syntax: Applying Objective Methods to Sentence Judgments , Newbury Park, CA: Sage Publications.
  • Cowie, Fiona, 1999, What’s Within? Nativism Reconsidered , New York: Oxford University Press.
  • Culicover, Peter W. and Jackendoff, Ray S., 2010, “Quantitative methods alone are not enough: Response to Gibson and Fedorenko”, Trends in Cognitive Science , 14(6): 234–235.
  • den Dikken, Marcel, Bernstein, Judy, Tortora, Christina, and Zanuttini, Raffaella, 2007, “Data and grammar: Means and individuals”, Theoretical Linguistics , 33: 269–318.
  • Derwing, Bruce, 1973, Transformational Grammar as a Theory of Language Acquisition: A Study in the Empirical, Conceptual and Methodological Foundations of Contemporary Linguistics , Cambridge: Cambridge University Press.
  • Devitt, Michael, 2006, Ignorance of Language , Oxford: Clarendon Press.
  • Dummett, Michael, 1986, “‘A nice derangement of epitaphs’: Some comments on Davidson and Hacking”, in Truth and Interpretation , Ernest Lepore (ed.), Oxford: Blackwell, 459–476.
  • Dunbar, Robin, 1998, “Theory of mind and the evolution of language”, in James R. Hurford, Michael Studdert-Kennedy, and Chris Knight (eds.), Approaches to the Evolution of Language , Cambridge: Cambridge University Press, 92–110.
  • Eckert, Penelope, 1989, Jocks and Burnouts: Social Categories and Identity in the High School , New York, NY: Teachers College Press.
  • Edelman, Shimon and Christiansen, Morten, 2003, “How seriously should we take minimalist syntax?”, Trends in Cognitive Sciences , 7: 60–61.
  • Elman, Jeffrey L., 1993, “Learning and development in neural networks: The importance of starting small”, Cognition , 48: 71–99.
  • Everett, Daniel, 2017, How language began: The story of humanity’s greatest invention , New York: W. W. Norton.
  • Ferreira, Fernanda, 2005, “Psycholinguistics, formal grammars, and cognitive science”, The Linguistic Review , 22: 365–380.
  • Field, Hartry, 1980, Science without Numbers: A Defense of Nominalism , Princeton, N.J.: Princeton University Press.
  • Fitch, W. Tecumseh, 2010, “Prolegomena to a future science of biolinguistics”, Biolinguistics , 3(4): 283–320.
  • Fodor, Janet Dean and Crowther, Carrie, 2002, “Understanding stimulus poverty arguments”, The Linguistic Review , 19(1–2): 105–146.
  • Fodor, Jerry A., 1983, The Modularity of Mind: An Essay on Faculty Psychology , Cambridge, Massachusetts: MIT Press.
  • Fodor, Jerry A. and Pylyshyn, Zenon W., 1988, “Connectionism and cognitive architecture: A critical analysis”, Cognition , 28: 3–71.
  • Foraker, Stephani, Regier, Terry, Khetarpal, Naveen, Perfors, Amy, and Tenenbaum, Joshua, 2009, “Indirect evidence and the poverty of the stimulus: The case of anaphoric one”, Cognitive Science , 33: 287–300.
  • Gasparri, Luca, 2021, “A pluralistic theory of wordhood”, Mind and Language , 36(4): 592–609.
  • George, Alexander, 1989, “How not to become confused about linguistics”, in Reflections on Chomsky , Alexander George (ed.), Oxford: Basil Blackwell, 90–110.
  • Gibson, Edward and Fedorenko, Evelina, 2013, “The need for quantitative methods in syntax and semantics research”, Language and Cognitive Processes , 28: 88–124. (See also the authors’ summary in Trends in Cognitive Sciences , 2010, Volume 14, pp. 233–234.)
  • Gold, E. Mark, 1967, “Language identification in the limit”, Information and Control , 10: 447–474.
  • Goldberg, Adele, 1995, Constructions: A Construction Grammar Approach to Argument Structure , Chicago: University of Chicago Press.
  • Gray, Russell D., 2005, “Pushing the time barrier in the quest for language roots”, Science , 209: 307–308.
  • Gray, Russell D., Drummond, A. J., and Greenhill, S. J., 2009, “Phylogenies reveal expansion pulses and pauses in Pacific settlement”, Science , 323: 479–483.
  • Gray, Russell D. and Jordan, Fiona M., 2000, “Language trees support the express-train sequence of Austronesian expansion”, Nature , 405: 1052–1055. Minor technical correction noted in Nature , 409: 743 (8 February 2001).
  • Greenhill, S. J., Atkinson, Q. D., Meade, A., and Gray, R. D., 2010, “The shape and tempo of language evolution”, Proceedings of the Royal Society B , 277: 2443–2450.
  • Harris, Zellig, 1957, “Co-occurrence and transformation in linguistic structure”, Language , 33: 283–340.
  • Haspelmath, Martin, 2011, “The indeterminacy of word segmentation and the nature of morphology and syntax”, Folia Linguistica , 45(1): 31–80.
  • Hauser, Marc D., Chomsky, Noam, and Fitch, W. Tecumseh, 2002 [HCF], “The faculty of language: What is it, who has it, and how did it evolve”, Science , 298: 1569–1579.
  • Hawthorne, John, and Lepore, Ernest, 2011, “On words”, The Journal of Philosophy , 108(9): 447–485.
  • Hellman, Geoffrey, 1989, Mathematics without Numbers: Towards a Modal-Structural Interpretation , Oxford: Clarendon Press.
  • Hockett, Charles F., 1968, The State of the Art , The Hague: Mouton.
  • Hoijer, Harry, 1954, “The Sapir Whorf hypothesis”, in Language in Culture , Harry Hoijer (ed.), Chicago: University of Chicago Press, 92–105.
  • Hopper, Paul and Thompson, Sandra, 1993, “Language universals, discourse pragmatics, and semantics”, Linguistic Sciences , 15: 357–376.
  • Horner, Victoria, Whiten, Andrew, Flynn, Emma, and de Waal, Frans B. M., 2006, “Faithful replication of foraging techniques along cultural transmission chains by chimpanzees and children”, Proceedings of the National Academy of Sciences , 103: 13878–13883.
  • Hurford, James R., 2000, “Social transmission favours linguistic generalization”, in J. R. Hurford, M. Studdert-Kennedy, and C. Knight (eds.), The Evolutionary Emergence of Language: Social Function and the Origins of Linguistic Form , Cambridge: Cambridge University Press, 219–230.
  • Irmak, Nurbay, 2019, “An ontology of words”, Erkenntnis , 84(5): 1139–1158.
  • Itkonen, Esa, 1978, Grammatical Theory and Metascience: A Critical Investigation into the Methodological and Philosophical Foundations of ‘Autonomous’ Linguistics , Berlin: Walter de Gruyter.
  • –––, 2013, “The philosophy of linguistics”, in Keith Allan (ed.), The Oxford Handbook of the History of Linguistics , Oxford: Oxford University Press, 747–774.
  • Jackendoff, Ray S., 2002, Foundations of Language: Brain, Meaning, Grammar, Evolution , Oxford: Oxford University Press.
  • –––, 2018, “Representations and Rules of Language”, in The Philosophy of Daniel Dennett , B. Huebner (ed.), Oxford University Press, 95–126.
  • Jacobson, Pauline, 1996, “The syntax/semantics interface in categorial grammar”, in Handbook of Contemporary Semantic Theory , Shalom Lappin (ed.), Cambridge, Massachusetts: Oxford University Press, 89–116.
  • Jain, Sanjay, Osherson, Daniel N., Royer, James S., and Sharma, Arun, 1999, Systems That Learn , Cambridge, Massachusetts: MIT Press, 2nd ed.
  • Johnson, Kent, 2004, “Gold’s theorem and cognitive science”, Philosophy of Science , 70(4): 571–592.
  • –––, 2007, “On the systematicity of language and thought”, Journal of Philosophy , 101: 111–139.
  • –––, 2015, “Notational variants and invariance in linguistics”, Mind and Language , 30(2): 162–186.
  • Joos, Martin (ed.), 1966, Readings in Linguistics I: The Development of Descriptive Linguistics in America 1925–56 , Chicago, IL: University of Chicago Press, fourth ed.
  • Kaplan, David, 1990, “Words”, Proceedings of the Aristotelian Society , 64: 93–119.
  • –––, 2011, “Words on words”, The Journal of Philosophy , 108(9): 504–529.
  • Karlsson, Fred, 2007, “Constraints on multiple center-embedding of clauses”, Journal of Linguistics , 43(2): 365–392.
  • Katz, Jerrold J., 1980, “Chomsky on meaning”, Language , 56(1): 1–41.
  • –––, 1981, Language and Other Abstract Objects , Totowa, NJ: Rowman and Littlefield.
  • ––– (ed.), 1985, Philosophy of Linguistics , Oxford: Oxford University Press.
  • –––, 1996, “The Unfinished Chomskyan Revolution”, Mind and Language , 11(3): 270–294.
  • –––, 1998, Realistic Rationalism , Cambridge, Massachusetts: MIT Press.
  • Kay, Paul and Regier, Terry, 2006, “Language, thought and color: Recent developments”, Trends in Cognitive Sciences , 10(2): 51–53.
  • Kay, Paul, Berlin, Brent, Maffi, Luisa, Merrifield, William, 2011, The World Color Survey , Stanford, Center for the Study of Language and Information.
  • Kirby, Simon, 2001, “Spontaneous evolution of linguistic structure: An iterated learning model of the emergence of regularity and irregularity”, IEEE Transactions on Evolutionary Computation , 5(2): 102–110.
  • Kirby, Simon, Cornish, Hannah, and Smith, Kenny, 2008, “Cumulative cultural evolution in the laboratory: An experimental approach to the origins of structure in human language”, Proceedings of the National Academy of Sciences , 101(31): 10681–10686.
  • Kousta, S. T., Vinson, D. P., and Vigliocco, G., 2008, “Investigating linguistic relativity through bilingualism: The case of grammatical gender”, Journal of Experimental Psychology: Learning, Memory, and Cognition , 34(4): 843–858.
  • Labov, William, 1966, The Social Stratification of English in New York City , Washington, DC: Center for Applied Linguistics. 2nd edition Cambridge University Press, 2006.
  • –––, 1975, “Empirical foundations of linguistic theory”, in The Scope of American Linguistics , R. Austerlitz (ed.), Lisse: Peter de Ridder, 77–133.
  • –––, 1996, “When intuitions fail”, in L. McNair, K. Singer, L. Dobrin, and M. Aucon (eds.), Papers from the Parasession on Theory and Data in Linguistics , Chicago: Chicago Linguistic Society, 77–106.
  • Lappin, Shalom and Shieber, Stuart, 2007, “Machine learning theory and practice as a source of insight into universal grammar”, Journal of Linguistics , 43: 393–427.
  • Larson, Richard, 1988, “On the double object construction”, Linguistic Inquiry , 19: 335–391.
  • Laurence, Stephen and Margolis, Eric, 2001, “The poverty of the stimulus argument”, British Journal of Philosophy of Science , 52(2): 217–276.
  • Levelt, W. J. M., 2008, An Introduction to the Theory of Formal Languages and Automata , Amsterdam: John Benjamins.
  • Lewis, David, 1969, Convention: A Philosophical Study , Cambridge, Massachusetts: Harvard University Press.
  • Li, P. and Gleitman, Lila, 2002, “Turning the tables: Language and spatial reasoning”, Cognition , 83: 265–294.
  • Lucy, John, 1996, “The scope of linguistic relativity: An analysis and review of empirical research”, in J. Gumperz and S. Levinson (eds.), Rethinking Linguistic Relativity , Cambridge: Cambridge University Press, 37–69.
  • Ludlow, Peter, 2011, The Philosophy of Generative Linguistics , Oxford: Oxford University Press.
  • MacWhinney, Brian, 2005, “The emergence of grammar from perspective taking”, in D. Pecher and R. A. Zwaan (eds.), The Grounding of Cognition , Cambridge: Cambridge University Press, 198–223.
  • Mallory, Fintan, 2020, “Linguistic types are capacity-individuated action-types”. Inquiry , 63(9–10): 1123–1148.
  • Martinet, André, 1960, Elements of General Linguistics , London: Faber.
  • Matthews, Robert, 1984, “The plausibility of rationalism”, Journal of Philosophy , 81: 492–515. Reprinted in Matthews and Demopoulos (1989), 51–75.
  • –––, 2007, “The case for linguistic nativism”, in Contemporary Debates in Cognitive Science , Robert J. Stainton, (ed.), Oxford: Blackwell, 81–96.
  • Matthews, Robert and Demopoulos, William (eds.), 1989, Learnability and Language Acquisition , Dordrecht: Foris.
  • Miller, James, 2020, “The ontology of words: Realism, nominalism, and eliminativism”, Philosophy Compass , 15(7): e12691.
  • –––, 2021, “A bundle theory of words”, Synthese , 198: 5731–5748.
  • Millikan, Ruth Garrett, 2003, “In defense of public language”, in L. M. Antony and N. Hornstein (eds.), Chomsky and His Critics , Oxford: Blackwell, 215–237.
  • Montague, Richard, 1974, Formal Philosophy: Selected Papers of Richard Montague , New Haven: Yale University Press. Edited by R. Thomason.
  • Morris, Charles, 1938, Foundations of the Theory of Signs , Chicago: University of Chicago Press.
  • Napoli, Donna Jo, 1996, Linguistics: An Introduction , New York, NY: Oxford University Press.
  • Nefdt, Ryan, 2016, “Languages and other abstract structures”, in Essays on Linguistic Realism , C. Behme and M. Neef (eds.), Amsterdam: John Benjamins Publishing, 139–184.
  • –––, 2019a, “The philosophy of linguistics: scientific underpinnings and methodological disputes”, Philosophy Compass , 14(12): e12636.
  • –––, 2019b, “The ontology of words: a structural approach”, Inquiry , 62(8): 877–911.
  • –––, 2019c, “Infinity and the foundations of linguistics”, Synthese , 196: 1671–1711.
  • –––, 2020, “Formal semantics and applied mathematics: an inferential account”, Journal of Logic, Language and Information , 29(2): 221–253.
  • –––, 2021, “Structural realism and generative linguistics”, Synthese , 199: 3711–3737.
  • Newmeyer, Frederick J., 1986, Linguistic Theory in America , New York: Academic Press, 2nd edition.
  • –––, 1991, “Functional explanation in linguistics and the origins of language”, Language and Communication , 11(1–2): 3–28.
  • –––, 1998, “On the supposed ‘counterfunctionality’ of universal grammar: Some evolutionary implications”, in J. R. Hurford, M. Studdert-Kennedy, and C. Knight (eds.), Approaches to the Evolution of Language , Cambridge: Cambridge University Press, 305–319.
  • –––, 2007, “Commentary on Sam Featherston, ‘Data in generative grammar: The stick and the carrot”’, Theoretical Linguistics , 33: 395–399.
  • Newport, Elissa L., 1988, “Constraints on learning and their role in language acquisition: Studies of the acquisition of American sign language”, Language Sciences , 10: 147–172.
  • O’Grady, William, 2008, “The emergentist program”, Lingua , 118: 447–464.
  • Osherson, Daniel N., Stob, Michael, and Weinstein, Scott, 1984, “Learning theory and natural language”, Cognition , 17(1): 1–28. Reprinted in Matthews and Demopoulos (1989), 19–50.
  • Partee, Barbara, 1975, “Montague grammar and transformational grammar”, Linguistic Inquiry , 6: 203–300.
  • Pelletier, Francis Jeffry, 1991, “The principle of semantic compositionality”, Topoi , 13: 11–24; reprinted, with additions, in S. Davis and B. Gillon, Semantics: A Reader , Oxford: Oxford University Press, 2004, pp. 133–156.
  • Penn, Julia, 1972, Linguistic Relativity versus Innate Ideas: The Origins of the Sapir-Whorf Hypothesis in German Thought , Paris: Mouton.
  • Phillips, Colin, 2010, “Should we impeach armchair linguists?”, in S. Iwasaki, H. Hoji, P. Clancy, and S.-O. Sohn (eds.), Japanese-Korean Linguistics 17 , Stanford, CA: CSLI Publications, 49–64.
  • Piattelli-Palmarini, Massimo, 1989, “Evolution, selection, and cognition: From ‘learning’ to parameter setting in biology and the study of language”, Cognition , 31: 1–44.
  • Pinker, Steven, 1994, The Language Instinct: The New Science of Language and Mind , New York, NY: Morrow Press.
  • –––, 2007, The Stuff of Thought: Language as a Window into Human Nature , New York, NY: Viking Penguin.
  • Pinker, Steven and Bloom, Paul, 1990, “Natural language and natural selection”, Behavioral and Brain Sciences , 13: 707–726.
  • Pinker, Steven and Jackendoff, Ray S., 2005, “The faculty of language: What’s special about it?”, Cognition , 95: 201–236.
  • Postal, Paul, 2003, “Remarks on the foundations of linguistics”, The Philosophical Forum , 34: 233–251.
  • Power, Camilla, 1998, “‘Old wives’ tales’: The gossip hypothesis and the reliability of cheap signals”, in J. R. Hurford, M. Studdert-Kennedy, and C. Knight (eds.), Approaches to the Evolution of Language , Cambridge: Cambridge University Press, 111–129.
  • Prinz, Jesse, 2002, Furnishing the Mind: Concepts and Their Perceptual Basis , Cambridge, Massachusetts: MIT Press.
  • Progovac, Ljiljana, 2015, Evolutionary syntax (Oxford Studies in the Evolution of Language), Oxford: Oxford University Press.
  • Pullum, Geoffrey K., 1983, “How many possible human languages are there?”, Linguistic Inquiry , 14: 447–467.
  • –––, 2013, “The central question in comparative syntactic metatheory”, Mind and Language , 28(4): 492–521.
  • –––, 2019, “Philosophy of linguistics”, in Kelly Michael Becker and Iain Thomson, (eds.), The Cambridge History of Philosophy, 1945–2015 , Cambridge: Cambridge University Press, 49–59.
  • Pullum, Geoffrey K. and Scholz, Barbara C., 1997, “Theoretical linguistics and the ontology of linguistic structure”, in T. Haukioja, M.-L. Helasvuo, and M. Miestamo, (eds.), SKY 1997: 1997 Yearbook of the Linguistic Association of Finland , Turku: Suomen kielitieteelinen yhdistys [Linguistic Association of Finland], 25–47.
  • –––, 2002, “Empirical assessment of stimulus poverty arguments”, The Linguistic Review , 19: 9–50.
  • –––, 2007, “ Systematicity and Natural Language Syntax”, Croatian Journal of Philosophy , 21: 375–402.
  • –––, 2010, “Recursion and the infinitude claim”, in Recursion in Human Language , Harry van der Hulst (ed.), Berlin: Mouton de Gruyter, no. 104 in Studies in Generative Grammar, 113–138.
  • Putnam, Hilary, 1963, “Probability and confirmation”, in The Voice of America Forum Lectures (Philosophy of Science Series, No. 10), Hilary Putnam, (ed.), Washington, D.C.: United States Information Agency. Reprinted in Mathematics, Matter and Method , Cambridge: Cambridge University Press, 1975, 293–304.
  • Quine, Willard Van Orman, 1972, “Linguistics and philosophy”, in Language and Philosophy: A Symposium , Sidney Hook (ed.), New York: New York University Press, 95–98.
  • –––, 1987, Quiddities: An Intermittently Philosophical Dictionary , Cambridge, Massachusetts: Harvard University Press.
  • Rey, Georges, 2006, “The intentional inexistence of language—But not cars”, in R. J. Stainton (ed.), Contemporary debates in cognitive science , Oxford: Blackwell, 237–255.
  • –––, 2020, Representation of Language: philosophical issues in a Chomskyan linguistics , Oxford: Oxford University Press.
  • Rohde, D. L. T. and Plaut, D. C., 1999, “Language acquisition in the absence of explicit negative evidence: How important is starting small?”, Cognition , 72: 67–109.
  • Roland, Doug and Jurafsky, Daniel, 2002, “Verb sense and verb subcategorization probabilities”, in S. Stevenson and P. Merlo (eds.), The Lexical Basis of Sentence Processing: Formal, Computational, and Experimental Issues , Amsterdam: John Benjamins, 325–346.
  • Ross, John R., 2010, “The Category Squish: Endstation Hauptwort”, Cognitive Linguistics Bibliography (CogBib) , Berlin, Boston: De Gruyter Mouton, 316–339.
  • Sampson, Geoffrey, 2001, Empirical Linguistics , London: Continuum Press.
  • –––, 2005, The Language Instinct Debate , London: Continuum Press.
  • Santana, Carlos, 2016, “What Is Language?” Ergo , 3(19): 501–523.
  • Sapir, Edward, 1921, Language , New York, NY: Harcourt.
  • –––, 1929, “The status of linguistics as a science”, Language , 5: 207–214. Reprinted in David Mandelbaum (ed.), Selected Writings of Edward Sapir in Language Culture and Personality , Berkeley and Los Angeles: University of California Press, 1968, 160–166.
  • Saussure, Ferdinand de, 1916, Cours de linguistique générale , Paris and Lausanne: Payot. Edited and published after Saussure’s death by Charles Bally and Albert Sechehaye with the collaboration of Albert Riedlinger. English translation by Roy Harris (1998) in Ferinand de Saussure Course in General Linguistics , New York: Open Court.
  • Scholz, Barbara C. and Pullum, Geoffrey K., 2002, “Searching for arguments to support linguistic nativism”, The Linguistic Review , 19: 185–223.
  • –––, 2006, “Irrational nativist exuberance”, in Contemporary Debates in Cognitive Science , Robert J. Stainton (ed.), Oxford: Basil Blackwell, 59–80.
  • –––, 2007, “Tracking the origins of generative grammar”, Journal of Linguistics , 43: 701–723.
  • Schütze, Carson, 1996, The Empirical Base of Linguistics: Grammaticality Judgments and Linguistic Methodology , Chicago: University of Chicago Press.
  • Seuren, Pieter A. M., 1998, Western Linguistics: An Historical Introduction , Oxford: Blackwell.
  • Shapiro, Stewart, 1997, Philosophy of Mathematics: Structure and Ontology , Oxford University Press.
  • Shinohara, Takeshi, 1990, “Inductive inference of monotonic formal systems from positive data”, in S. Arikawa, S. Goto, S. Ohsuga, and T. Yokomori (eds.), Algorithmic Learning Theory , Berlin: Springer, 339–351.
  • Skyrms, Brian, 2010, Signals: Evolution, Learning and Information , Oxford: Oxford University Press.
  • Slobin, Dan, 1996, “From thought and language to thinking for speaking”, in J. Gumperz and S. Levinson (eds.), Rethinking Linguistic Relativity , Cambridge: Cambridge University Press, 70–96.
  • Soames, Scott, 1984, “Linguistics and psychology”, Linguistics and Philosophy , 7: 155–179.
  • Sperber, Dan, and Origgi, Gloria, 2010, “A pragmatic perspective on the evolution of language”, in R. K. Larson, V. Déprez, and H. Yamakido (eds.), The Evolution of Human Language: Biolinguistic Perspectives , Cambridge: Cambridge University Press, 124–132.
  • Sprouse, Jon, 2011, “A test of the cognitive assumptions of magnitude estimation: Commutativity does not hold for acceptability judgments”, Language , 87(2): 274–288.
  • Sprouse, Jon and Almeida, Diogo, 2012, “Assessing the reliability of textbook data in syntax: Adger’s Core Syntax ”, Journal of Linguistics , 48(3): 609–652.
  • Stainton, Robert J. (ed.), 2006, Contemporary Debates in Cognitive Science , Oxford: Blackwell.
  • Stainton, Robert J., 2014, “Philosophy of Linguistics”. Oxford Handbooks Online , published online July 2014. doi:10.1093/oxfordhb/9780199935314.013.002
  • Steedman, Mark, 2000, The Syntactic Process , Cambridge, Massachusetts: MIT Press.
  • –––, 2017, “The emergence of language”, Mind and Language , 32(5): 597–590.
  • Szabó, Zoltan, 1999, “Expressions and their representation”, The Philosophical Quarterly , 49(195): 145–163.
  • –––, 2015, “Major parts of speech”, Erkenntnis , 80: 3–29.
  • Szabolcsi, Anna, 1997, “Strategies for scope taking”, in Ways of Scope Taking , Anna Szabolcsi (ed.), Dordrecht: Kluwer, 109–155.
  • Thierry, Guillaume, Athanasopulous, Panos, Wiggett, Alison, Dering, Benjamin, and Kuipers, Jan-Rouke, 2009, “Unconscious effects of language-specific terminology on pre-attentive color perception”, Proceedings of the National Academy of Sciences , 106(11): 4567–4570.
  • Tomalin, Marcus, 2006, Linguistics and the Formal Sciences: The Origins of Generative Grammar , Cambridge: Cambridge University Press.
  • Tomasello, Michael, 1998, “Introduction”, in The New Psychology of Language: Cognitive and Functional Approaches to Language Structure , Michael Tomasello (ed.), Mahwah, NJ: Lawrence Erlbaum.
  • –––, 2003, Constructing a Language: A Usage-Based Theory of Language Acquisition , Cambridge, MA: Harvard University Press.
  • –––, 2008, Origins of Human Communication , Cambridge, MA: Bradford Books/MIT Press.
  • Tomlin, Russell S., 1990, “Functionalism in second language acquisition”, Studies in Second Language Acquisition , 12: 155–177.
  • Valian, Virginia, 1982, “Psycholinguistic experiment and linguistic intuition”, in T. W. Simon and R. J. Scholes (eds.), Language, Mind, and Brain , Hillsdale, NJ: Lawrence Erlbaum, 179–188.
  • Van Valin, Robert, 1991, “Functionalist linguistic theory and language acquisition”, First Language , 11: 7–40.
  • Voegelin, Carl F. and Harris, Zellig S., 1951, “Methods for determining intelligibility among dialects of natural languages”, Proceedings of the American Philosophical Society , 95(3): 322–329.
  • Wasow, Thomas and Arnold, Jennifer, 2005, “Intuitions in linguistic argumentation”, Lingua , 115: 1481–1496.
  • Weinreich, Max, 1945, “Der yivo un di problemen fun undzer tsayt”, Yivo Bleter , 25: 3–18.
  • Weskott, Thomas and Fanselow, Gisbert, 2011, “On the informativity of different measures of linguistic acceptability”, Language , 87(2): 249–273.
  • Weisberg, Michael, 2013, Simulation and Similarity: Using Models to Understand the World , New York: Oxford University Press.
  • Wetzel, Linda, 2009, Types and tokens: An essay on abstract objects , Boston, MA: MIT Press.
  • Wexler, Kenneth and Culicover, Peter, 1980, Formal Principles of Language Acquisition , Cambridge, Massachusetts: MIT Press.
  • Wexler, Kenneth and Hamburger, Henry, 1973, “On the insufficiency of surface data for the learning of transformational languages”, in J. Hintikka, J. Moravcsik, and P. Suppes (eds.), Approaches to Natural Language , Dordrecht: Reidel, 16–179.
  • Whorf, Benjamin Lee, 1956, Language, Thought and Reality , Cambridge University Press: MIT Press. Edited by John B. Carroll.
  • Worden, Robert, 1998, “The evolution of language from social intelligence”, in J. R. Hurford, M. Studdert-Kennedy, and C. Knight (eds.), Approaches to the Evolution of Language , Cambridge: Cambridge University Press, 148–166.
How to cite this entry . Preview the PDF version of this entry at the Friends of the SEP Society . Look up topics and thinkers related to this entry at the Internet Philosophy Ontology Project (InPhO). Enhanced bibliography for this entry at PhilPapers , with links to its database.
  • Treebank , Wikipedia.
  • The Switchboard Corpus .
  • The British National Corpus .

analysis | assertion | compositionality | defaults in semantics and pragmatics | descriptions | empiricism: logical | idiolects | innate/acquired distinction | innateness: and language | language of thought hypothesis | linguistics: computational | logic: intensional | mental representation | pragmatics | propositional attitude reports | reference | relativism | rigid designators

Acknowledgments

The authors are very grateful to the two SEP referees, Tom Wasow and William Starr, who provided careful reviews of our drafts; to Bonnie Webber and Zoltan Galsi for insightful comments and advice; and to Dean Mellow for some helpful corrections. BCS was the lead author of this article throughout the lengthy period of its preparation, and worked on it in collaboration with FJP and GKP until the post-refereeing revision was submitted at the end of April 2011. She died two weeks later, on May 14. FJP and GKP oversaw the few final corrections that were made when the HTML version was first published in September 2011.

Copyright © 2024 by Barbara C. Scholz Francis Jeffry Pelletier < francisp @ ualberta . ca > Geoffrey K. Pullum < pullum @ gmail . com > Ryan Nefdt < ryan . nefdt @ uct . ac . za >

  • Accessibility

Support SEP

Mirror sites.

View this site from another server:

  • Info about mirror sites

The Stanford Encyclopedia of Philosophy is copyright © 2024 by The Metaphysics Research Lab , Department of Philosophy, Stanford University

Library of Congress Catalog Data: ISSN 1095-5054

  • Search Menu
  • Browse content in Arts and Humanities
  • Browse content in Archaeology
  • Anglo-Saxon and Medieval Archaeology
  • Archaeological Methodology and Techniques
  • Archaeology by Region
  • Archaeology of Religion
  • Archaeology of Trade and Exchange
  • Biblical Archaeology
  • Contemporary and Public Archaeology
  • Environmental Archaeology
  • Historical Archaeology
  • History and Theory of Archaeology
  • Industrial Archaeology
  • Landscape Archaeology
  • Mortuary Archaeology
  • Prehistoric Archaeology
  • Underwater Archaeology
  • Urban Archaeology
  • Zooarchaeology
  • Browse content in Architecture
  • Architectural Structure and Design
  • History of Architecture
  • Residential and Domestic Buildings
  • Theory of Architecture
  • Browse content in Art
  • Art Subjects and Themes
  • History of Art
  • Industrial and Commercial Art
  • Theory of Art
  • Biographical Studies
  • Byzantine Studies
  • Browse content in Classical Studies
  • Classical History
  • Classical Philosophy
  • Classical Mythology
  • Classical Literature
  • Classical Reception
  • Classical Art and Architecture
  • Classical Oratory and Rhetoric
  • Greek and Roman Papyrology
  • Greek and Roman Epigraphy
  • Greek and Roman Law
  • Greek and Roman Archaeology
  • Late Antiquity
  • Religion in the Ancient World
  • Digital Humanities
  • Browse content in History
  • Colonialism and Imperialism
  • Diplomatic History
  • Environmental History
  • Genealogy, Heraldry, Names, and Honours
  • Genocide and Ethnic Cleansing
  • Historical Geography
  • History by Period
  • History of Emotions
  • History of Agriculture
  • History of Education
  • History of Gender and Sexuality
  • Industrial History
  • Intellectual History
  • International History
  • Labour History
  • Legal and Constitutional History
  • Local and Family History
  • Maritime History
  • Military History
  • National Liberation and Post-Colonialism
  • Oral History
  • Political History
  • Public History
  • Regional and National History
  • Revolutions and Rebellions
  • Slavery and Abolition of Slavery
  • Social and Cultural History
  • Theory, Methods, and Historiography
  • Urban History
  • World History
  • Browse content in Language Teaching and Learning
  • Language Learning (Specific Skills)
  • Language Teaching Theory and Methods
  • Browse content in Linguistics
  • Applied Linguistics
  • Cognitive Linguistics
  • Computational Linguistics
  • Forensic Linguistics
  • Grammar, Syntax and Morphology
  • Historical and Diachronic Linguistics
  • History of English
  • Language Evolution
  • Language Reference
  • Language Acquisition
  • Language Variation
  • Language Families
  • Lexicography
  • Linguistic Anthropology
  • Linguistic Theories
  • Linguistic Typology
  • Phonetics and Phonology
  • Psycholinguistics
  • Sociolinguistics
  • Translation and Interpretation
  • Writing Systems
  • Browse content in Literature
  • Bibliography
  • Children's Literature Studies
  • Literary Studies (Romanticism)
  • Literary Studies (American)
  • Literary Studies (Asian)
  • Literary Studies (European)
  • Literary Studies (Eco-criticism)
  • Literary Studies (Modernism)
  • Literary Studies - World
  • Literary Studies (1500 to 1800)
  • Literary Studies (19th Century)
  • Literary Studies (20th Century onwards)
  • Literary Studies (African American Literature)
  • Literary Studies (British and Irish)
  • Literary Studies (Early and Medieval)
  • Literary Studies (Fiction, Novelists, and Prose Writers)
  • Literary Studies (Gender Studies)
  • Literary Studies (Graphic Novels)
  • Literary Studies (History of the Book)
  • Literary Studies (Plays and Playwrights)
  • Literary Studies (Poetry and Poets)
  • Literary Studies (Postcolonial Literature)
  • Literary Studies (Queer Studies)
  • Literary Studies (Science Fiction)
  • Literary Studies (Travel Literature)
  • Literary Studies (War Literature)
  • Literary Studies (Women's Writing)
  • Literary Theory and Cultural Studies
  • Mythology and Folklore
  • Shakespeare Studies and Criticism
  • Browse content in Media Studies
  • Browse content in Music
  • Applied Music
  • Dance and Music
  • Ethics in Music
  • Ethnomusicology
  • Gender and Sexuality in Music
  • Medicine and Music
  • Music Cultures
  • Music and Media
  • Music and Religion
  • Music and Culture
  • Music Education and Pedagogy
  • Music Theory and Analysis
  • Musical Scores, Lyrics, and Libretti
  • Musical Structures, Styles, and Techniques
  • Musicology and Music History
  • Performance Practice and Studies
  • Race and Ethnicity in Music
  • Sound Studies
  • Browse content in Performing Arts
  • Browse content in Philosophy
  • Aesthetics and Philosophy of Art
  • Epistemology
  • Feminist Philosophy
  • History of Western Philosophy
  • Metaphysics
  • Moral Philosophy
  • Non-Western Philosophy
  • Philosophy of Language
  • Philosophy of Mind
  • Philosophy of Perception
  • Philosophy of Science
  • Philosophy of Action
  • Philosophy of Law
  • Philosophy of Religion
  • Philosophy of Mathematics and Logic
  • Practical Ethics
  • Social and Political Philosophy
  • Browse content in Religion
  • Biblical Studies
  • Christianity
  • East Asian Religions
  • History of Religion
  • Judaism and Jewish Studies
  • Qumran Studies
  • Religion and Education
  • Religion and Health
  • Religion and Politics
  • Religion and Science
  • Religion and Law
  • Religion and Art, Literature, and Music
  • Religious Studies
  • Browse content in Society and Culture
  • Cookery, Food, and Drink
  • Cultural Studies
  • Customs and Traditions
  • Ethical Issues and Debates
  • Hobbies, Games, Arts and Crafts
  • Lifestyle, Home, and Garden
  • Natural world, Country Life, and Pets
  • Popular Beliefs and Controversial Knowledge
  • Sports and Outdoor Recreation
  • Technology and Society
  • Travel and Holiday
  • Visual Culture
  • Browse content in Law
  • Arbitration
  • Browse content in Company and Commercial Law
  • Commercial Law
  • Company Law
  • Browse content in Comparative Law
  • Systems of Law
  • Competition Law
  • Browse content in Constitutional and Administrative Law
  • Government Powers
  • Judicial Review
  • Local Government Law
  • Military and Defence Law
  • Parliamentary and Legislative Practice
  • Construction Law
  • Contract Law
  • Browse content in Criminal Law
  • Criminal Procedure
  • Criminal Evidence Law
  • Sentencing and Punishment
  • Employment and Labour Law
  • Environment and Energy Law
  • Browse content in Financial Law
  • Banking Law
  • Insolvency Law
  • History of Law
  • Human Rights and Immigration
  • Intellectual Property Law
  • Browse content in International Law
  • Private International Law and Conflict of Laws
  • Public International Law
  • IT and Communications Law
  • Jurisprudence and Philosophy of Law
  • Law and Politics
  • Law and Society
  • Browse content in Legal System and Practice
  • Courts and Procedure
  • Legal Skills and Practice
  • Primary Sources of Law
  • Regulation of Legal Profession
  • Medical and Healthcare Law
  • Browse content in Policing
  • Criminal Investigation and Detection
  • Police and Security Services
  • Police Procedure and Law
  • Police Regional Planning
  • Browse content in Property Law
  • Personal Property Law
  • Study and Revision
  • Terrorism and National Security Law
  • Browse content in Trusts Law
  • Wills and Probate or Succession
  • Browse content in Medicine and Health
  • Browse content in Allied Health Professions
  • Arts Therapies
  • Clinical Science
  • Dietetics and Nutrition
  • Occupational Therapy
  • Operating Department Practice
  • Physiotherapy
  • Radiography
  • Speech and Language Therapy
  • Browse content in Anaesthetics
  • General Anaesthesia
  • Neuroanaesthesia
  • Clinical Neuroscience
  • Browse content in Clinical Medicine
  • Acute Medicine
  • Cardiovascular Medicine
  • Clinical Genetics
  • Clinical Pharmacology and Therapeutics
  • Dermatology
  • Endocrinology and Diabetes
  • Gastroenterology
  • Genito-urinary Medicine
  • Geriatric Medicine
  • Infectious Diseases
  • Medical Toxicology
  • Medical Oncology
  • Pain Medicine
  • Palliative Medicine
  • Rehabilitation Medicine
  • Respiratory Medicine and Pulmonology
  • Rheumatology
  • Sleep Medicine
  • Sports and Exercise Medicine
  • Community Medical Services
  • Critical Care
  • Emergency Medicine
  • Forensic Medicine
  • Haematology
  • History of Medicine
  • Browse content in Medical Skills
  • Clinical Skills
  • Communication Skills
  • Nursing Skills
  • Surgical Skills
  • Browse content in Medical Dentistry
  • Oral and Maxillofacial Surgery
  • Paediatric Dentistry
  • Restorative Dentistry and Orthodontics
  • Surgical Dentistry
  • Medical Ethics
  • Medical Statistics and Methodology
  • Browse content in Neurology
  • Clinical Neurophysiology
  • Neuropathology
  • Nursing Studies
  • Browse content in Obstetrics and Gynaecology
  • Gynaecology
  • Occupational Medicine
  • Ophthalmology
  • Otolaryngology (ENT)
  • Browse content in Paediatrics
  • Neonatology
  • Browse content in Pathology
  • Chemical Pathology
  • Clinical Cytogenetics and Molecular Genetics
  • Histopathology
  • Medical Microbiology and Virology
  • Patient Education and Information
  • Browse content in Pharmacology
  • Psychopharmacology
  • Browse content in Popular Health
  • Caring for Others
  • Complementary and Alternative Medicine
  • Self-help and Personal Development
  • Browse content in Preclinical Medicine
  • Cell Biology
  • Molecular Biology and Genetics
  • Reproduction, Growth and Development
  • Primary Care
  • Professional Development in Medicine
  • Browse content in Psychiatry
  • Addiction Medicine
  • Child and Adolescent Psychiatry
  • Forensic Psychiatry
  • Learning Disabilities
  • Old Age Psychiatry
  • Psychotherapy
  • Browse content in Public Health and Epidemiology
  • Epidemiology
  • Public Health
  • Browse content in Radiology
  • Clinical Radiology
  • Interventional Radiology
  • Nuclear Medicine
  • Radiation Oncology
  • Reproductive Medicine
  • Browse content in Surgery
  • Cardiothoracic Surgery
  • Gastro-intestinal and Colorectal Surgery
  • General Surgery
  • Neurosurgery
  • Paediatric Surgery
  • Peri-operative Care
  • Plastic and Reconstructive Surgery
  • Surgical Oncology
  • Transplant Surgery
  • Trauma and Orthopaedic Surgery
  • Vascular Surgery
  • Browse content in Science and Mathematics
  • Browse content in Biological Sciences
  • Aquatic Biology
  • Biochemistry
  • Bioinformatics and Computational Biology
  • Developmental Biology
  • Ecology and Conservation
  • Evolutionary Biology
  • Genetics and Genomics
  • Microbiology
  • Molecular and Cell Biology
  • Natural History
  • Plant Sciences and Forestry
  • Research Methods in Life Sciences
  • Structural Biology
  • Systems Biology
  • Zoology and Animal Sciences
  • Browse content in Chemistry
  • Analytical Chemistry
  • Computational Chemistry
  • Crystallography
  • Environmental Chemistry
  • Industrial Chemistry
  • Inorganic Chemistry
  • Materials Chemistry
  • Medicinal Chemistry
  • Mineralogy and Gems
  • Organic Chemistry
  • Physical Chemistry
  • Polymer Chemistry
  • Study and Communication Skills in Chemistry
  • Theoretical Chemistry
  • Browse content in Computer Science
  • Artificial Intelligence
  • Computer Architecture and Logic Design
  • Game Studies
  • Human-Computer Interaction
  • Mathematical Theory of Computation
  • Programming Languages
  • Software Engineering
  • Systems Analysis and Design
  • Virtual Reality
  • Browse content in Computing
  • Business Applications
  • Computer Security
  • Computer Games
  • Computer Networking and Communications
  • Digital Lifestyle
  • Graphical and Digital Media Applications
  • Operating Systems
  • Browse content in Earth Sciences and Geography
  • Atmospheric Sciences
  • Environmental Geography
  • Geology and the Lithosphere
  • Maps and Map-making
  • Meteorology and Climatology
  • Oceanography and Hydrology
  • Palaeontology
  • Physical Geography and Topography
  • Regional Geography
  • Soil Science
  • Urban Geography
  • Browse content in Engineering and Technology
  • Agriculture and Farming
  • Biological Engineering
  • Civil Engineering, Surveying, and Building
  • Electronics and Communications Engineering
  • Energy Technology
  • Engineering (General)
  • Environmental Science, Engineering, and Technology
  • History of Engineering and Technology
  • Mechanical Engineering and Materials
  • Technology of Industrial Chemistry
  • Transport Technology and Trades
  • Browse content in Environmental Science
  • Applied Ecology (Environmental Science)
  • Conservation of the Environment (Environmental Science)
  • Environmental Sustainability
  • Environmentalist Thought and Ideology (Environmental Science)
  • Management of Land and Natural Resources (Environmental Science)
  • Natural Disasters (Environmental Science)
  • Nuclear Issues (Environmental Science)
  • Pollution and Threats to the Environment (Environmental Science)
  • Social Impact of Environmental Issues (Environmental Science)
  • History of Science and Technology
  • Browse content in Materials Science
  • Ceramics and Glasses
  • Composite Materials
  • Metals, Alloying, and Corrosion
  • Nanotechnology
  • Browse content in Mathematics
  • Applied Mathematics
  • Biomathematics and Statistics
  • History of Mathematics
  • Mathematical Education
  • Mathematical Finance
  • Mathematical Analysis
  • Numerical and Computational Mathematics
  • Probability and Statistics
  • Pure Mathematics
  • Browse content in Neuroscience
  • Cognition and Behavioural Neuroscience
  • Development of the Nervous System
  • Disorders of the Nervous System
  • History of Neuroscience
  • Invertebrate Neurobiology
  • Molecular and Cellular Systems
  • Neuroendocrinology and Autonomic Nervous System
  • Neuroscientific Techniques
  • Sensory and Motor Systems
  • Browse content in Physics
  • Astronomy and Astrophysics
  • Atomic, Molecular, and Optical Physics
  • Biological and Medical Physics
  • Classical Mechanics
  • Computational Physics
  • Condensed Matter Physics
  • Electromagnetism, Optics, and Acoustics
  • History of Physics
  • Mathematical and Statistical Physics
  • Measurement Science
  • Nuclear Physics
  • Particles and Fields
  • Plasma Physics
  • Quantum Physics
  • Relativity and Gravitation
  • Semiconductor and Mesoscopic Physics
  • Browse content in Psychology
  • Affective Sciences
  • Clinical Psychology
  • Cognitive Psychology
  • Cognitive Neuroscience
  • Criminal and Forensic Psychology
  • Developmental Psychology
  • Educational Psychology
  • Evolutionary Psychology
  • Health Psychology
  • History and Systems in Psychology
  • Music Psychology
  • Neuropsychology
  • Organizational Psychology
  • Psychological Assessment and Testing
  • Psychology of Human-Technology Interaction
  • Psychology Professional Development and Training
  • Research Methods in Psychology
  • Social Psychology
  • Browse content in Social Sciences
  • Browse content in Anthropology
  • Anthropology of Religion
  • Human Evolution
  • Medical Anthropology
  • Physical Anthropology
  • Regional Anthropology
  • Social and Cultural Anthropology
  • Theory and Practice of Anthropology
  • Browse content in Business and Management
  • Business Ethics
  • Business Strategy
  • Business History
  • Business and Technology
  • Business and Government
  • Business and the Environment
  • Comparative Management
  • Corporate Governance
  • Corporate Social Responsibility
  • Entrepreneurship
  • Health Management
  • Human Resource Management
  • Industrial and Employment Relations
  • Industry Studies
  • Information and Communication Technologies
  • International Business
  • Knowledge Management
  • Management and Management Techniques
  • Operations Management
  • Organizational Theory and Behaviour
  • Pensions and Pension Management
  • Public and Nonprofit Management
  • Strategic Management
  • Supply Chain Management
  • Browse content in Criminology and Criminal Justice
  • Criminal Justice
  • Criminology
  • Forms of Crime
  • International and Comparative Criminology
  • Youth Violence and Juvenile Justice
  • Development Studies
  • Browse content in Economics
  • Agricultural, Environmental, and Natural Resource Economics
  • Asian Economics
  • Behavioural Finance
  • Behavioural Economics and Neuroeconomics
  • Econometrics and Mathematical Economics
  • Economic History
  • Economic Systems
  • Economic Methodology
  • Economic Development and Growth
  • Financial Markets
  • Financial Institutions and Services
  • General Economics and Teaching
  • Health, Education, and Welfare
  • History of Economic Thought
  • International Economics
  • Labour and Demographic Economics
  • Law and Economics
  • Macroeconomics and Monetary Economics
  • Microeconomics
  • Public Economics
  • Urban, Rural, and Regional Economics
  • Welfare Economics
  • Browse content in Education
  • Adult Education and Continuous Learning
  • Care and Counselling of Students
  • Early Childhood and Elementary Education
  • Educational Equipment and Technology
  • Educational Strategies and Policy
  • Higher and Further Education
  • Organization and Management of Education
  • Philosophy and Theory of Education
  • Schools Studies
  • Secondary Education
  • Teaching of a Specific Subject
  • Teaching of Specific Groups and Special Educational Needs
  • Teaching Skills and Techniques
  • Browse content in Environment
  • Applied Ecology (Social Science)
  • Climate Change
  • Conservation of the Environment (Social Science)
  • Environmentalist Thought and Ideology (Social Science)
  • Natural Disasters (Environment)
  • Social Impact of Environmental Issues (Social Science)
  • Browse content in Human Geography
  • Cultural Geography
  • Economic Geography
  • Political Geography
  • Browse content in Interdisciplinary Studies
  • Communication Studies
  • Museums, Libraries, and Information Sciences
  • Browse content in Politics
  • African Politics
  • Asian Politics
  • Chinese Politics
  • Comparative Politics
  • Conflict Politics
  • Elections and Electoral Studies
  • Environmental Politics
  • European Union
  • Foreign Policy
  • Gender and Politics
  • Human Rights and Politics
  • Indian Politics
  • International Relations
  • International Organization (Politics)
  • International Political Economy
  • Irish Politics
  • Latin American Politics
  • Middle Eastern Politics
  • Political Behaviour
  • Political Economy
  • Political Institutions
  • Political Methodology
  • Political Communication
  • Political Philosophy
  • Political Sociology
  • Political Theory
  • Politics and Law
  • Public Policy
  • Public Administration
  • Quantitative Political Methodology
  • Regional Political Studies
  • Russian Politics
  • Security Studies
  • State and Local Government
  • UK Politics
  • US Politics
  • Browse content in Regional and Area Studies
  • African Studies
  • Asian Studies
  • East Asian Studies
  • Japanese Studies
  • Latin American Studies
  • Middle Eastern Studies
  • Native American Studies
  • Scottish Studies
  • Browse content in Research and Information
  • Research Methods
  • Browse content in Social Work
  • Addictions and Substance Misuse
  • Adoption and Fostering
  • Care of the Elderly
  • Child and Adolescent Social Work
  • Couple and Family Social Work
  • Developmental and Physical Disabilities Social Work
  • Direct Practice and Clinical Social Work
  • Emergency Services
  • Human Behaviour and the Social Environment
  • International and Global Issues in Social Work
  • Mental and Behavioural Health
  • Social Justice and Human Rights
  • Social Policy and Advocacy
  • Social Work and Crime and Justice
  • Social Work Macro Practice
  • Social Work Practice Settings
  • Social Work Research and Evidence-based Practice
  • Welfare and Benefit Systems
  • Browse content in Sociology
  • Childhood Studies
  • Community Development
  • Comparative and Historical Sociology
  • Economic Sociology
  • Gender and Sexuality
  • Gerontology and Ageing
  • Health, Illness, and Medicine
  • Marriage and the Family
  • Migration Studies
  • Occupations, Professions, and Work
  • Organizations
  • Population and Demography
  • Race and Ethnicity
  • Social Theory
  • Social Movements and Social Change
  • Social Research and Statistics
  • Social Stratification, Inequality, and Mobility
  • Sociology of Religion
  • Sociology of Education
  • Sport and Leisure
  • Urban and Rural Studies
  • Browse content in Warfare and Defence
  • Defence Strategy, Planning, and Research
  • Land Forces and Warfare
  • Military Administration
  • Military Life and Institutions
  • Naval Forces and Warfare
  • Other Warfare and Defence Issues
  • Peace Studies and Conflict Resolution
  • Weapons and Equipment

The Oxford Handbook of Cognitive Linguistics

  • < Previous chapter
  • Next chapter >

38 Cognitive Linguistics and Linguistic Relativity

Eric Pederson (PhD 1991) is associate professor of Linguistics at the University of Oregon. The overarching theme of his research is the relationship between language and conceptual processes. He was a student at the University of California, Berkeley, working within Cognitive Linguistics with George Lakoff, Dan Slobin, Eve Sweetser, and Leonard Talmy since 1980. He joined the Max Planck Institute for Psycholinguistics in 1991 until 1997, where he began working on issues more specific to linguistic relativity. Relevant publications include “Geographic and Manipulable Space in Two Tamil Linguistic Systems” (1993); “Language as Context, Language as Means: Spatial Cognition and Habitual Language use” (1995); “Semantic Typology and Spatial Conceptualization” (with Eve Danziger, Stephen Levinson, Sotaro Kita, Gunter Senft, and David Wilkins, 1998); “Through the Looking Glass: Literacy, Writing Systems and Mirror Image Discrimination” (with Eve Danziger, 1998); and “Mirror-Image Discrimination among Nonliterate, Monoliterate, and Biliterate Tamil Speakers” (2003). In addition to linguistic relativity, his general interests include semantic typology, field/descriptive linguistics (South India), and the representation of events. Eric Pederson can be reached at [email protected].

  • Published: 18 September 2012
  • Cite Icon Cite
  • Permissions Icon Permissions

Linguistic relativity (also known as the Sapir-Whorf Hypothesis) is a general cover term for the conjunction of two basic notions. The first notion is that languages are relative, that is, that they vary in their expression of concepts in noteworthy ways. The second notion is that the linguistic expression of concepts has some degree of influence over conceptualization in cognitive domains, which need not necessarily be linguistically mediated. This article explores the treatment of linguistic relativity within works generally representative of cognitive linguistics and presents a survey of classic and more modern (pre- and post-1980s) research within linguistics, anthropology, and psychology. First, it provides a brief overview of the history of linguistic relativity theorizing from Wilhelm von Humboldt through to Benjamin Whorf. It then discusses the role of literacy to cognitive and cultural development, folk classification, and formulations of linguistic relativity.

1. Introduction

Linguistic relativity (also known as the Sapir - Whorf Hypothesis ) is a general cover term for the conjunction of two basic notions. The first notion is that languages are relative , that is, that they vary in their expression of concepts in noteworthy ways. What constitutes “noteworthy” is, of course, a matter of some interpretation. Cognitive scientists interested in human universals will often describe some particular linguistic variation as essentially minor, while others, for example, some anthropological linguists, may describe the same variation as significant.

The second component notion to linguistic relativity is that the linguistic expression of concepts has some degree of influence over conceptualization in cognitive domains, which need not necessarily be linguistically mediated. In textbooks, this notion of language affecting conceptualization is typically divided into “strong” and “weak” hypotheses. The “strong” hypothesis (also known as linguistic determinism ) is that the variable categories of language essentially control the available categories of general cognition. As thus stated, this “strong” hypothesis is typically dismissed as untenable. The “weak” hypothesis states that the linguistic categories may influence the categories of thought but are not fundamentally restrictive. As thus stated, this “weak” hypothesis is typically considered trivially true.

Arguably, this simplification of the broad issue of the relationship between linguistic and cognitive categorization into two simple (“strong” vs. “weak”) statements has impeded development of genuinely testable hypotheses and has helped lead studies of linguistic relativity into academic ill-repute. Modern research into the general question of linguistic relativity has focused on more narrowly stated hypotheses for testing, that is, investigating the specific relationships between particular linguistic categories (e.g., the categories of number, color, or spatial direction) and more exactly specified cognitive operations (e.g., encoding into long-term memory or deductive reasoning).

This chapter is organized as (i) a brief history of the research question (section 2 ); (ii) a discussion of the challenges in designing research into linguistic relativity (section 3 ); (iii) the treatment of linguistic relativity within works generally representative of Cognitive Linguistics (section 4 ); and (iv) a survey of classic and more modern (pre- and post-1980s) research within linguistics, anthropology, and psychology (section 5 ).

In addition to this chapter, several other surveys of linguistic relativity may be consulted. Lucy ( 1997a ) gives a broad overview of different approaches which have investigated linguistic relativity, while Lucy ( 1992b ) elaborates on a particular empirical approach and provides detailed critiques of previous empirical work. Lee ( 1996 ) provides historical documentation to the often poorly understood work of Benjamin Lee Whorf (see also Lee 2000 ). Hill and Mannheim ( 1992 ) trace the history of the notion of world view with respect to language through twentieth-century anthropology, from Boas through Cognitive Linguistics of the 1980s to the work of John Lucy. Hill and Mannheim also provides a useful overview of the anthropological cum semiotic approach to culturally embedded language use—see especially Hanks ( 1990 ) and Silverstein ( 1985 , 1987 ).

Smith ( 1996 ) also discusses the writings of Sapir and Whorf to clarify that most popular accounts of the Sapir-Whorf Hypothesis are not directly derivative of their work. She is also concerned that the relatively large-scale dismissal of the Sapir-Whorf Hypothesis in academic culture has been at the expense of serious research into the relationships between language and thought. Similar discussion of the “demise” of the “Whorf Hypothesis” and the misconstrual of Whorf's actual writings can be found in Alford ( 1978 ). 1

Koerner ( 2000 ) also provides a survey of the “pedigree” of linguistic relativity “from Locke to Lucy,” that is, from the seventeenth through the twentieth century. chapters 10–12 of Foley ( 1997 ) as well provide historical coverage of the notion, with summaries of fairly recent work with spatial language and classifiers. Duranti ( 1997 ) similarly provides historical coverage with particular emphasis on the American anthropology traditions.

Hunt and Agnoli ( 1991 ) revisit linguistic relativity from the perspective of cognitive psychology, which had largely rejected the notion as either false or uninteresting during the 1970s. Within canonical Cognitive Linguistics, Lakoff ( 1987 ) dedicates chapter 18 of Women , Fire , and Dangerous Things to discussions of evidence for and types of linguistic relativity. Many of the principles from that chapter have informed the remainder of his work.

2. Historical Speculation and Modern Formulations

Given the wealth of historical surveys of linguistic relativity, this chapter will focus more on modern work and methodological issues. However, a brief overview of the history of linguistic relativity theorizing will help to situate the modern research questions.

2.1. From Humboldt through Whorf

The most widely cited intellectual antecedent for linguistic relativity is the work of Humboldt. Later, the work of Boas is widely seen as the inheritor of the Humboldtian notions and through him, the concern with linguistic relativity was taken up in the writings of Sapir, who developed the vital notion of the “patterns” or structural systematicity of language as being particularly relevant to the relationship between language, mind, and culture.

Humboldt's principal work addressing linguistic relativity is Über die Verschiedenheit des menschlichen Sprachbaues und ihren Einfluss auf die geistige Entwicklung des Menschengeschlecht [On the diversity of human language construction and its influence on the mental development of the human species]. There are many editions and translations of this work; for a recent edition of Peter Heath's English translation, see Losonsky ( 1999 ). The philosophical precursors to Humboldt, as well as linguistic relativity in general, is discussed in Manchester ( 1985 ), and an overview of Humboldt's notion of language and Weltansicht (‘world view’) is provided in Brown ( 1967 ).

The writings of Benjamin Lee Whorf are best known through Carroll's edited collection Whorf ( 1956 ). This collection helped to popularize the notion that the categories of language may influence the categories of thought. However, Lee ( 1996 ) argues—especially in light of the previously unpublished “Yale report” (see Whorf and Trager [ 1938 ] 1996 )—that Whorf was concerned with the interpenetration of language and thought; that is, the two words language and thought refer to aspects of a single system, and it is a misapprehension to ask in what way one affects the other. This is quite distinct from the more modular view of language processing dominant in current psychology and linguistics.

2.2. Literacy

While modern linguistics places considerable emphasis on spoken language—which means that this chapter will focus on the potential cognitive impact of the categories found in spoken or signed languages—the role of literacy to cognitive and cultural development has long been a subject of debate.

Early twentieth-century experiments on the relationship between literacy and cognitive development were conducted by Aleksandr Luria and colleagues (for an overview in English, see Luria 1976 ). This classic work investigated the effects of previously established, Soviet-era adult literacy programs on the development of various cognitive skills. There were a number of methodological problems with that work—perhaps the most significant one being the confounding of formal schooling with the acquisition of literacy (or conversely, the lack of formal schooling with nonliterate populations). The largest single effort to overcome this common confound is reported by Scribner and Cole ( 1981 ), who investigated effects of literacy acquisition in the absence of formal schooling. The designs and subject pools were still not completely free of confounding factors and the results, while fascinating, give a largely mixed picture of the effects of literacy as an independent factor on cognition.

“The literacy hypothesis,” namely that various cultural features can be traced to the development of literacy in the history of a given culture, has been subject to considerable debate. Goody and Watt ( 1962 ), one of the better known works, extolled the effects of specifically alphabetic literacy as critical in the development of early Greek and later European culture. This view came under considerable criticism, and Goody himself later backed away from the specific claims about alphabetic literacy. 2 However, on a more general level, the claim that literacy engenders certain cognitive changes—especially enhanced metalinguistic awareness—continues to be argued. Readers interested in the effects of literacy on cognition could also consult Scinto ( 1986 ), Graff ( 1987 ), Olson ( 1991 , 2002 ), Ong ( 1992 ), and references therein.

Rather than studying the general effects of reading and writing on cognition, one line of research has been concerned with the effects of learning particular writing systems. Morais et al. ( 1979 ) investigate the effects of child-acquired literacy on phonemic awareness, and Read et al. ( 1986 ) present evidence arguing that alphabetic literacy, but not logographic and syllabic literacy, leads to phonemic awareness. In Danziger and Pederson ( 1998 ) and Pederson ( 2003 ), I argue that familiarity with specific graphemic qualities can lead to differences in visual categorization in nonwriting/nonreading tasks.

2.3. Folk Classification

Anthropologists have long been concerned with folk classification , that is, the culturally specific ways in which linguistic and other categories are organized into coherent systems. Perhaps the richest body of work is in the area of taxonomies of natural kinds (plants, animals, etc.). This research is conveniently served by having a scientific standard for comparison. While there is abundant anecdotal evidence that people interact with natural kinds according to their taxonomical relations to other natural kinds (e.g., X is a pet, so treat it like other pets), there has not been much in the way of psychological-style testing of specific linguistic relativity hypotheses in this domain. For an introduction to folk classification, see Hunn ( 1977 , 1982 ), Berlin, Breedlove, and Raven ( 1973 ), Berlin ( 1978 ), and Blount ( 1993 ).

2.4. Formulations of Linguistic Relativity

There are many semantic domains one could search for linguistic relativity effects—that is, domains in which one might find linguistic categories conditioning nonlinguistic categorization. For example, cultures and languages are notorious for having varying kinship terms, which group into major types with various subtypes. Importantly, the categories of allowable behaviors with kin tend to correspond to the grouping by kinship terminology. For example, South Indian (Dravidian) languages systematically distinguish between cross-cousins and parallel cousins, with marriage allowed between cross-cousins and incest taboo applying to parallel cousins. In contrast, North Indian languages typically classify all cousins with siblings and incest taboo applies to all (see Carter 1973 ).

However important sexual reproduction may be to our species, the standards of marriage are clearly the result of cultural convention overlaid on biological predispositions. Accordingly, finding linguistic variation corresponding to categories of human behavior in such a domain is not generally taken as a particularly revealing demonstration of linguistic relativity. Likewise, elaborated vocabulary sets in expert domains and impoverished sets where there is little experience, however interesting, are also not taken as particularly revealing. While a tropical language speaker may lack the broad vocabulary of English for discussing frozen precipitation, that same speaker may be quite particular in distinguishing what English speakers lump together as ‘cousins’.

In other words, cases of categorization which are dependent on environmentally or culturally variable experience are generally considered uninteresting domains for the study of linguistic relativity. This corresponds to the late twentieth-century bias toward universalism in the cognitive sciences; namely, for variation to be noteworthy, it should be in a domain where variation was not previously thought to be possible. That is to say, for linguistic relativity to be broadly interesting, it must apply within cognitive domains which operate on “basic” and universal human experience.

3. Challenges in Researching Linguistic Relativity

3.1. intralinguistic variation.

Speakers may use language differently across different contexts, and this difference may be indicative of shifting conceptual representations. One of the few studies within Cognitive Linguistics to empirically address intralinguistic variation is Geeraerts, Grondelaers, and Bakema ( 1994 , especially chapter 4 : “Onomasiological Variation”), which explores alternative expressions as the representation of different construals and perspectivization.

Of course, some of these alternative expressions may be confined to some subcommunities and dialects. While linguistic relativity is typically discussed as the difference across speakers of distinct languages, there is every reason to wonder about parallels with differences in conceptualization that may exist within a single language community. Speakers of different dialects may have different linguistic patterns which might be hypothesized to correspond to different habitual conceptualizations. In Pederson ( 1993 , 1995 ), I investigate communities of Tamil speakers who systematically vary in their preference for terms of spatial reference, but who otherwise speak essentially the same dialect.

The work of Loftus ( 1975 ) has demonstrated that the choice of particular linguistic expressions at the time of encoding or recall may well influence nonlinguistic representation of events. Extrapolating from Loftus's work, we might wonder to what extent language generally can prime specific nonlinguistic representations—I call this the language as prime model. The fact that social humans are surrounded by linguistic input suggests that there might be a cumulative effect of this language priming. Indeed, if a particular linguistic encoding presented before a certain perception influences the nonlinguistic encoding or recall of that perception, what then might be the cumulative effect of one type of linguistic encoding rather than another being used throughout a speaker's personal history? If, for example, the classifiers of a speaker's habitual language force categorization of certain objects as ‘long and thin’, it seems reasonable that such objects may be remembered as potentially longer or thinner than they actually were.

Of course, if there were no consistent pattern to the linguistic priming, then we would not expect any single representation to become dominant. Indeed, Kay ( 1996 ) has argued that there is considerable flexibility within any language for alternative representations, and speakers may well alternate from one representation to another. This suggests that rather than a single and simple “world-view” necessary for a cleanly testable hypothesis, speakers may draw on complex “repertoires” of representations. While this does not preclude the possibility of systematic differences across languages having different repertoires, it certainly argues that the differences are far less obvious.

Given flexibility within a single language, a linguistic relativity hypothesis to be tested may need to compare patterns which are pervasive in one language and underexpressed in another language. This can be difficult to compensate for in an experimental design. A balanced design might seek opposing, but functionally equivalent systems, which are dominant in each language community. Each community may have both systems in common, but not to the same level of default familiarity. Of course, the experimental measure needs to be sufficiently non-priming itself so as to allow each subject population to rely on their default mode of representation.

3.2. Selecting a Domain

Universals in categorization may be of more than one type. Most relevantly, some categories may be essentially innate, that is, an internal predisposition of the organism. Other universal categories maybe the result of commonalities of all human environments in conjunction with our innately driven mechanisms. Even assuming that we can reliably presume that certain categories are universal, determining which are purely innate and which derive from interaction with universal properties of the environment is not a trivial task.

Variation in innate properties is impossible—except inasmuch as the variation is within innately proscribed limits—so we cannot look for linguistic relativity effects in these domains. For linguistic relativity effects to be both interesting to cognitive scientists and robust in their operations, they must apply in a domain which is generally presumed universal by virtue of the common environment, but which can be hypothesized to be nonuniversal. As discussed above, demonstrating effects from language type in cognitive domains with wide variation is unexciting. It follows that the researcher interested in testing linguistic relativity best seeks a domain which is hypothesized to be fairly basic to cognition, but just shy of exhibiting a universal pattern.

This motivates modern linguistic relativity studies to examine categorization in domains presumed to derive somewhat immediately from basic perceptual stimuli or fundamental mechanisms of reasoning. The majority of such empirical studies concern categorization of visual or spatial properties of objects or the environment. A few studies have examined purported differences in reasoning, but these are inherently more difficult to pursue. Object properties and the environment can be experimentally controlled, but processes of reasoning—especially in cross-cultural work—are notoriously difficult to measure while maintaining adequate control of subject variables.

3.3. Independent Evidence for Language and Cognition

Linguists—especially cognitive linguists—frequently claim that a particular linguistic form represents a particular underlying conceptualization. Obviously, however, any substantial claim of a relationship between language and cognition needs independent assessment of each and a correlation established between the two.

Perhaps surprisingly, most work on linguistic relativity spends remarkably little effort demonstrating the linguistic facts prior to seeking the hypothesized cognitive variable. Some of the most severe criticisms of linguistic relativity studies have worried about this insufficient linguistic description. Lucy ( 1992b ) is especially clear in his call for more careful linguistic analysis preparatory to linguistic relativity experimentation.

Given the relative accessibility of the linguistic facts compared with the difficulty inferring cognitive behavior from behavioral measures, one could argue that the often minimal characterization of language is of unacceptable sloppiness. More charitably, linguistic facts are typically quite complex, and in an effort to seek a testable hypothesis, a certain amount of simplification becomes inevitable. Unfortunately, there is no standard to use in evaluating the adequacy of a linguistic description for linguistic relativity work other than using the general standards of descriptive linguistics. Descriptive linguistics tends to be as exhaustive as is practically possible and does not necessarily foster the creation of simple hypotheses about linguistic and conceptual categorization. On the other hand, it is difficult to argue that studies in linguistic relativity should hold their linguistic descriptions to a lower standard.

A related problem is the variability of language. Since many different varieties of language exist depending on communicative and descriptive context, it can be quite misleading to speak of Hopi or English as having a specific characteristic, unless one can argue that this characteristic is true and uniquely true (e.g., there are no competitive constructions) in all contexts. This is, needless to say, a difficult endeavor, but failing to argue the general applicability of the pattern invites the next linguist with expertise in the language to pull forth numerous counterexamples. Studies most closely following the approaches advocated by Whorf have tended to focus on basic grammatical features of the language which are presumed to be fairly context independent. However, this may overlook other linguistic features which may well be relevant to a particular hypothesis of linguistic and conceptual categorization.

One way to partially circumvent this problem was followed in Pederson et al. ( 1998 ), which seeks to describe language characteristics typically used for, in this case, table-top spatial reference. There is no attempt to include or exclude information on the basis of whether or not the relevant language elements were grammaticized or lexicalized. Rather, if the information was present in the language used for a particular context, these linguistic categories are presumed to be available conceptual categories within same or similar contexts. This approach leaves unanswered the question of how broadly the linguistic description (or for that matter the cognitive description as well) applies to the subject population in a variety of other contexts, but it does help ensure that the linguistic description is the most exact match for the cognitive enquiry.

3.4. Subvocalization or What Is Nonlinguistic?

If independent measures are to be taken of both language use and cognitive processes, then great care is necessary to ensure that the behavioral measure for the nonlinguistic cognitive process is not covertly measuring linguistically mediated behavior.

Ideally, the entire cognitive task would be nonlinguistic, but as a practical minimum, the instructions and training for the task must be couched in language which is neutral with respect to the current hypothesis. This is particularly difficult to manage when a language has grammatically obligatory encoding. How do we interpret an effect which may be due to obligatory encoding in the instructions? Is this just an effect of the instructions, or can we interpret this as a general language effect because the instructions only exemplify the continual linguistic context the subjects live within?

There is a general presumption that instructions to the subjects should be in the subjects' native language. One might be tempted to use a shared second language as a type of neutral metalanguage for task instructions, but this introduces unexplored variables. If there is the possibility of a cognitive effect from the regular use of one's native language, then there is also the possibility of an effect from the immediate use of the language of instruction. Additionally, it is more difficult to be certain that all subjects understand the second-language instructions in exactly the same way as the experimenter. Finally, it is unclear how one would guarantee that the language of instruction is neutral with respect to anticipated behavioral outcomes. The very fact that it may mark different categories from the native language may influence the outcome in unpredictable ways.

It is safest therefore to minimize any language-based instruction. General instructions (e.g., “Sit here”) cannot be excluded, but critical information is best presented through neutral examples with minimal accompanying language. Since a dearth of talking makes it more difficult to monitor subject comprehension, it is imperative that the experimental design include a built-in check (e.g., control trials ) to ensure that each subject understands the task in the same way—except, of course, for the variation for which the task was designed to test. An account of the effects of subtle changes in instruction with children in explorations with base ten number systems can be found in Saxton and Towse ( 1998 ).

Another concern is that subjects involved in an ostensibly nonlanguage measure actually choose to use language as part of the means of determining their behavior. For example, the subjects may subvocalize their reasoning in a complex problem and then any patterning of behavior along the lines of the linguistic categories is scarcely surprising. In Pederson ( 1995 ), I address this concern by arguing that if subjects have distinct levels of linguistic and conceptual representations, they should only choose to approach a nonlinguistic task using linguistic means if there were a sufficiently close match between these two levels with respect to the experiment. In effect, a subject's unforced decision to rely on linguistic categories can be understood as validation of at least one sort of linguistic relativity hypothesis.

3.5. Finding Behavioral Consequences of Linguistically Determined Cognitive Variation

Variation in categorization of spatial or perceptual features can be of relatively minor consequence. Whether one thinks of pencils more fundamentally as tools or as long skinny objects has probably little effect on their employment.

The most basic features of humans and their environment are stable across linguistic communities. Gravity pulls in a constant direction, visual perception is roughly comparable, and so forth. If there are cognitive differences across communities with respect to universal features, then these different cognitive patterns must have functional equivalence ; that is, different ways of thinking about the same thing must largely allow the same behavioral responses. For example, whether a line of objects is understood as proceeding from left to right or from north to south makes little difference under most circumstances. If the objects are removed and the subject must rebuild them, either understanding of the array will give the same rebuilt line with no effect on accuracy. Accordingly, any experimental task must select an uncommon condition where the principle of functional equivalence fails to hold (see especially Levinson 1996 ). To continue this example, if the subject is rotated by 90 or 180 degrees before being asked to rebuild an array, the underlying representation (left-right or north-south) should result in a different direction for the rebuilding.

Without a context which effectively disambiguates the possible underlying representations from behavioral responses, a researcher must demonstrate that one subject population has a deficient or improved performance on a task and that this differential performance corresponds to a difference in (default) linguistic encoding. There is a long and sordid history of attributing deficiencies to populations that the investigator does not belong to. Accordingly, it is entirely appropriate that the burden of proof fall particularly hard on the researcher claiming that a studied population is somehow impaired on a given task as a result of their pattern of linguistic encoding. Even if the population is claimed to have an ability which is augmented by linguistic encoding, it is difficult to demonstrate that any difference in ability derives specifically from linguistic differences and not from any of a myriad of environmental (perhaps even nutritional) conditions.

Related to this is the concern for the ecological validity of the experimental task. A task may fail to measure subject ability or preferences owing to unfamiliarity of the materials, instructions, or testing context. Further, it is difficult to decide on the basis of just a few experiments which effects can be generalized to hold for nonexperimental contexts—to wit, the complexity of daily life. This is not, however, an argument against experimentation as the inherently interpretive nature of simple observational data ultimately requires experimentally controlled measures.

3.6. Types of Experimental Design

Various types of experimental tasks have been used for investigating the cognitive side of linguistic relativity. Whatever research methods are used, reliability of the results is far more likely if there is triangulation from a number of observational and experimental methods.

Sorting and Triads Tasks

Perhaps the most common design used in linguistic relativity studies is a sorting task. Quite simply, the subject is presented with a number of stimuli and is asked to group them into categories. These categories may be ad hoc (subject determined) or preselected (researcher determined). Multiple strategies may be used for the sorting task, giving different sorting results. The most common variant of the sorting task is the triads task which presents a single stimulus to the subjects and asks them to group it with either of two other stimuli or stimuli sets; that is, does stimulus X group better with A or with B? (hence, the term AXB test in some research paradigms). For an archetypal example of a triads task, see Davies et al. ( 1998 ).

This task is easy to administer as long as the stimuli are reasonably tangible, interpretable, and able to be considered in a nearly simultaneous manner. One consideration of sorting designs is that subjects often report awareness of multiple strategies which might be employed. Of course, the researcher cannot indicate which is a preferred strategy and can only instruct the subject to sort according to “first impression,” “whatever seems most natural,” or other such instructions. The interpretation of these instructions may add an uncontrolled variable. Further, sorting tasks inherently invite the subjects to respond according to their beliefs about the researcher's expectations, which may not in fact be what would be the normal sorting decision outside of this task.

Discrimination Tasks

Other tasks seek to find different discriminations across populations. As a practical consequence, differences usually boil down to one population making finer or more distinctions than another population; see, for example, much of the work on color discrimination and linguistic labeling discussed in the debates in Hardin and Maffi ( 1997 ). However, it is at least theoretically possible that one population might be more sensitive to certain features at the expense of other features and that a contrasting population would show the reverse pattern.

A limitation of discrimination tasks is that for them to be interpretable, one must be able to assume that beyond the independent variable of different linguistic systems, all subjects brought the same degree of attention, general task satisfying abilities, and so on to the experimental task. Should, for example, one population be less likely to be attentionally engaged, then this reduces the possibility of isolating a linguistic effect on cognition.

Problem Solving Tasks

Problem solving tasks are readily used in many types of research. In linguistic relativity studies, they are typically of two design types: difficult solution or alternative solution.

The first type involves a task which provides some difficulty in finding the solution. Some subjects are anticipated to be better or worse than others at solving the task. As with reduced discrimination just discussed, it is extremely difficult to argue that it is specifically the categories of language which lead to differential performance. The counterfactual reasoning task employed by Bloom ( 1981 ) was such a task, and the difficulty in interpreting its results was part of much of the controversy surrounding that work.

The second type of problem solving tasks allow for alternative solutions each of which should be indicative of a different underlying representation. As such, these are similar to triads tasks in that they allow each subject to find the most “natural” solution for them (at least within the given experimental context). For example, in Pederson ( 1995 ) I describe a transitivity task in which subjects know how each of two objects are spatially related to a third object. They must then decide which side of the second object the first/test object must be placed. Depending on how these relationships are encoded, the test object will be placed on a different side of the second object. Like triads tasks, there is the potential problem that the subjects may be aware of the possibility of multiple solutions, prompting responses derived from any number of uncontrolled factors.

Embedded Tasks

Within psychological research, there is a common solution to the problem of subject awareness of multiple possible responses. Namely, the actual measure of the task is embedded within another task for which the subject is more consciously aware. For example, subjects may be asked to respond as to whether a figure is masculine or feminine, but the researcher is really measuring the distribution of attention to the figures. While the embedded task may still be influenced by subject expectations, it is an indirect and presumably nonreflected influence. As such, one can argue that the responses measured by the embedded task are more likely to correspond to default behaviors used outside of this exact experimental context. The “Animals in a Row” task discussed in Pederson et al. ( 1998 ) was one such task, where subjects understood the task as one to recreate a sequence of toy animals, but the critical dependent measure was the direction the animals were facing when subjects placed them on the table-top before them.

Variable Responses

The researcher must also be careful in coding fixed response types from the subjects. It may be that subject preference is for a response type not allowed by the forced choice, and when pigeonholed into a different response type, subjects may not be responding in a manner reflecting their typical underlying representations. Also, certain patterns (or lack of patterns) of responses may actually indicate a preference for a response type not anticipated by the experimental design. For example, in the “Animals in a Row” task just discussed, some populations—and not others—appear on the scoring sheets as preserving the orientation of the original stimuli roughly half the time. On closer inspection, many of these subjects were actually entirely consistent in giving the animals the same orientation (e.g., always facing left) regardless of the original orientation of the stimuli. Since the task appeared to be about the order and not the orientation of the animals, this is a perfectly reasonable response. Unfortunately, there was no hypothesis anticipating this response, and no claims could be made as to why some subjects and not others gave this response pattern.

3.7. Controlling Extraneous Variables

Work such as Kay and Kempton ( 1984 ) demonstrates that the effects of native language on nonlinguistic categorization tasks can vary with even slightly varied task demands. This is commonly interpreted as an indication that “relativity effects” are “weak.” A more conservative interpretation is that there are many factors (of undetermined “strength”) which can effect results and that language may be only one of many possible factors. The exact total effect of language will depend on what other nonlinguistic factors are in effect. This requires that an experimental design for linguistic relativity effects carefully control all foreseeable linguistic and nonlinguistic variables.

Linguistic Variables

Since they are most directly related to the tested hypothesis, language variables are perhaps the most critical to control in one's design.

Of fundamental importance is that one must be certain that the base language of the subjects is consistent with respect to whatever features have led to the specific hypothesis. This may seem trivial, but dialectal (and even idiolectal) variation may well have the effect that some speakers do not share certain critical linguistic features even though they ostensibly speak the same language.

Perhaps even more problematic is the issue of bilingualism. Unless all subjects are totally monolingual, this is a potential problem for the design. Generally, linguistic relativity tests presume that one's “native” language capacity is the most relevant, but this cannot preclude effects from other known languages. Age of acquisition of second languages may also vary widely; there is certainly no established model of the effects of age of acquisition on nonlinguistic category formation.

If nonnative categories have been learned, how can we assume that they are not also brought to bear on the experimental task—clouding the results in unpredictable ways? This is perhaps most insidious when the language of instruction differs from the native language. Suitably, then, serious work in linguistic relativity needs to use the native language for instruction, but even this is not necessarily a straightforward task. For example, how does one ensure that instructions to multiple populations are both exactly and suitably translated?

How to Control for Exact Translations in a Comparative Work?

Work in linguistic relativity has had an impact in translation theory. Indeed, belief in a sufficiently strong model of insurmountable language differences would suggest that complete translations would be difficult to attain. House ( 2000 ) presents an overview of the challenges of translation and suggests a solution to the problem of linguistic relativity and translation. Chafe ( 2000 ) also discusses translation issues with respect to linguistic relativity, and Slobin ( 1991 , 1996 ) uses translations in his discussions of how languages most suitably express motion events (see the section on space, below). The work of Bloom and his critics (see the discussion below) is particularly relevant for this issue because the ability to translate the experimental task from English to Chinese was central to his research question of counterfactual reasoning. Indeed, one might be skeptical of any attempt to investigate linguistic relativity in which the nonlinguistic experimental design is essentially a language-based task.

Of immediate practical concern is the translation of instructions for any research instrument itself. It is difficult enough to be confident that two subjects speaking the same language have the same understanding of a task's instructions. How, then, can the researcher be confident that translations of instructions are understood identically by speakers of different languages especially in the context of an experiment which seeks to confirm that speakers of these different languages in fact do understand the world in different ways?

The most obvious solution is to avoid linguistic instruction entirely. This does not remove the possibility that subjects understand the task differently, but it does ensure that any different understanding is not the direct result of immediate linguistic context. However, there are severe restrictions on what can be reliably and efficiently instructed without language. Understandably, then, most research relies on language-based instruction. In such cases, one must seek to phrase instructions in such a way that one sample is not more influenced by the particular choice of phrasing than the other sample.

To invent an example, imagine we are interested in the effect of evidential marking (linguistic markings which indicate how information is known to the speaker) on the salience of sources of even nonlinguistic information to speakers of a language which obligatorily marks evidentiality. This population would contrast with speakers of a language which essentially lacks routine marking. How, then, might we word our instructions? Do we use expressions typical for each language such that one set of instructions contains evidential marking and the other not? Alternatively, do we provide evidential information for both languages? In the case of the language which does not typically mark evidentials, providing this information would obviously be more “marked” in usage than for the other language. This greater markedness of the information might make the evidential information more salient for those subjects who normally do not concern themselves with any language expression of evidentiality, which in turn could make issues of evidentiality more salient than they would be under average conditions—countering the entire design of the experiment!

Recent Language Use

Another potential language factor affecting results might be preexperimental, but recent, language use. If the language of instruction can influence results, could not language use immediately prior to instruction also influence the results? Indeed, if we assume that linguistic categories prime access to parallel nonlinguistic categories, then how do we control for language use outside of the experimental setting? On the one hand, one could argue that language use outside of the experiment is exactly the independent variable under consideration, and this is controlled simply through subject selection. On the other hand, if a language has multiple ways of representing categories, what is the potential effect if a subject has most recently been using one of the less typical linguistic categories for his or her language? Once again, the cleanest solution to this risk is to test categories for which there is minimal linguistic variation within each of the examined languages. 3

Conversation during Task

The last of the language variables to consider is language use during the experiment itself. Lucy and Shwedder ( 1988 ) found that forbidding subjects to have conversations between exposure and recall in a memory task allowed a greater recall of focal color terms than of nonfocal color terms (see the subsection on color below). Subjects who had (unrelated) conversations remembered focal and nonfocal colors about equally well. While Lucy and Shwedder do not provide a model for why this might be the case, it clearly suggests that even incidental language use during and perhaps around a task can have significant influences on performance. Other work (see Gennari et al. 2002 ) has suggested that even in cases where there might normally be no particular relation between habitual language use and performance on a nonlinguistic task, language used during exposure or memorization to stimuli can lead to nonlinguistic responses in alignment with language use.

Nonlinguistic Subject Variables

Even more heterogonous to a subject sample than the linguistic variables are the cultural, educational, and other experiential variables. Subject questionnaires are the usual ways to try to control these variables in post hoc analysis, but this control is limited by the foresight to collect adequate information.

One of the more obvious variables to control or record is the amount of schooling and literacy. Unfortunately, while schooling is easily represented on an ordinal scale (first to postsecondary grades), there is little guarantee that this represents the same education especially across, but even within, two population samples. For example, literacy is also not as simple a variable as it might appear. Subjects may be literate in different languages (and scripts) and may have very different literacy practices. Coding subjects who only read the Bible in their nonnative language and other subjects who read a variety of materials in their native language as both simply “literate” clearly glosses over potentially significant differences in experience.

Expertise may also vary considerably across samples. One of the most thorny obstacles in cross-cultural psychology is comparing testing results across two populations, one of which habitually engages with experiment-like settings and the other of which does not. This may have effects beyond simple difficulty in performance, but may affect the way in which subjects understand instructions, second-guess the intentions of the experimenter, and so on. 4

Sex or gender, age, and the more physiologically based experiences are also difficult to compare. Being a woman in different societies means very different daily experiences beyond the variables of amount of schooling and the like. To what extent are subjects in their thirties the same across two populations. In one society but not another, a 35-year-old might typically be a grandparent in declining health with uncorrected vision or hearing loss.

Testing Environment

Lastly, variation in the testing environment is often difficult to control. The more broadly cross-cultural the samplings, the greater the dependence on local conditions. One might think of the ideal as an identical laboratory setup for each population sampled. However, since different subjects might react differently within such an environment, this is not necessarily a panacea (in addition to the obvious practical difficulty in implementation).

The best approach is to carefully examine the environmental features needed for the task at hand. If an experiment is about color categorization, lighting obviously needs to be controlled; if an experiment is about spatial arrays, adjacent landmarks and handedness need to be controlled; and so on. For example, in the basic experiment reported in Pederson et al. ( 1998 ), the use of table tops was not considered essential for tasks testing “table-top space,” but the use of two delimited testing surfaces and the geometrical relationship and distances between these surfaces was critical to the design. This allowed the individual experimenters to set up tables or mats on the ground/floor as was more appropriate for the broader material culture. 5

3.8. Establishing Causal Directionality

Once a correlation between a language pattern and a behavioral response has been experimentally established, the problem of establishing causal directionality remains. While this is a problem for any correlational design, it is particularly vexing for studies of linguistic relativity. Quite simply, it is difficult to rule out the possibility that subjects habitually speak the way they do as a consequence of their culture (and environment) as opposed to the possibility that the culture thinks the way it does because of their language. For discussions of the role of culture vis-à-vis language in linguistic relativity studies, see Bickel ( 2000 ), Enfield ( 2000 ), and the fairly standard reference of Hanks ( 1990 ).

In specific response to work on spatial cognition, Li and Gleitman ( 2002 ) argue that behavioral response patterns are not causally attributable to community language preferences, but rather that language use reflects cultural practice and concerns, for example, the many words for snow used by skiers—however, see also Levinson et al. ( 2002 ) for an extensive response. To the extent that the language features under investigation are roughly as changeable as the culture, this is certainly a likely possibility. On the other hand, when the language features are essentially fossilized in the grammatical system, they cannot be understood as the consequences of current cultural conditions. If anything, the pattern of grammaticized distinctions reflects the fossilized conceptualizations of one's ancestors.

4. Work within Cognitive Linguistics

Some of the earliest cognitive linguistic work (1970s) explicitly tying grammatical structure to cognition is found in studies by Talmy (see especially Talmy 1977 , 1978 ). This work largely focuses on the universal (or at least broadly found) patterns of language and has been revised and expanded in Talmy ( 2000a , 2000b ). Talmy treats language as one of many “cognitive systems” which has the “set of grammatically specified notions [constitute] the fundamental conceptual structuring system of language.… Thus, grammar broadly conceived, is the determinant of conceptual structure within one cognitive system, language” (2000a: 21–22). However, the relationship between this cognitive system (language) and others (i.e., nonlinguistic cognition) is relatively unspecified in his work. Structural commonalities between the various cognitive systems are suggested—most specifically between visual perception and language—but, importantly, Talmy avoids claims that there is any causal effect from linguistic categories to nonlinguistic categories. 6

Langacker is bolder in the relationship between grammar and cognition: in Cognitive Grammar's “view of linguistic semantics. Meaning is equated with conceptualization (in the broadest sense)” (Langacker 1987 : 55). Langacker ( 1991 ) further argues that the cognitive models underlying clause structure have prototypes which are rooted in (variable) cultural understanding. To the extent that we find interesting cross-linguistic variation, we can see the work of Talmy and Langacker as sources for linguistic relativity hypotheses to test—as, for example, Slobin ( 1996 , 2000 ) has begun with the motion event typology of Talmy ( 1985 ).

As mentioned above, Lakoff ( 1987 : chapter 18 ) directly addresses linguistic relativity. Within this chapter on linguistic relativity, there is a discussion of different ways in which two cross-linguistic systems might be “commensurate.” They might be translatable , understandable (though this is vaguely defined), commensurate in usage, share the same framing , and/or use the same organization of the various underlying concepts. In addition to a summary of the now classic Kay and Kempton ( 1984 ), there is an elaborate extension to linguistic relativity of semantics work in Mixtec and English by Brugman ( 1981 ) and Brugman and Macaulay ( 1986 ).

Metaphor is an obvious area of interest to many cognitive linguists (see Grady, this volume, chapter 8 , and references therein). The nature of metaphor is to consider conceptualizations in terms of other linguistically expressed domains. To the extent that source domains can vary cross-linguistically or cross-culturally (or different features of these source domains are mapped), this is an area ripe for linguistic relativity studies. To date, however, linguistic relativity studies—that is to say, work with behavioral data—have largely limited themselves to the study of elemental and literal language. One exception to this is linguistic relativity research on time, which almost necessarily is metaphorically expressed (see section 5.6 below).

5. Research by Topic Area

This section gives a brief overview of modern linguistic relativity work organized by topic area. While some comments are given, it is impossible in this space to summarize the findings of the entire body of work. Further, the empirical details of each study are essential to critical evaluation of the findings, so the original sources must be consulted.

Perhaps the greatest debate in linguistic relativity has been in the domain of color. Historically, linguists and anthropologists had been struck by the seemingly boundless diversity in color nomenclature. Given the obvious biological underpinnings of color perception, this made “color” a domain of choice to seek language-specific effects overriding biological prerequisites.

Lenneberg and Roberts ( 1956 ) is one of the earliest attempts to empirically test linguistic relativity, and as such this study spends considerable space defining the intellectual concerns before it reports on a relatively small study involving Zuni versus English color categorization. Brown and Lenneberg ( 1958 ) report on various work and develop the notion of codability : that is, the use of language as a way to more efficient coding of categories for the purposes not only of communication, but also of augmenting personal memory.

Berlin and Kay ( 1969 ) and the updated methodology in Kay and McDaniel ( 1978 ) have laid the groundwork of considerable research in color terminology. Central to the method is the use of Munsell color chips as a reference standard which can be carried to various field sites. Universal patterns were found to establish a typology of different color systems which appeared to be built out of a small set of universal principles. Research continues to be robust in this area and the interested reader may wish to consult the conference proceedings published as Hardin and Maffi ( 1997 ) for more current perspectives.

Eleanor Rosch (under her previous name: Heider 1971 , 1972 ) found that focal colors (or Hering primaries from Hering's theory of light and color, see Hering 1964 ) were better remembered even by young children and were also more perceptually salient for them. Further, Heider and Olivier ( 1972 ) and Rosch ( 1973 ) found that, even for members of a community (the Dani of Papua New Guinea) who had little color terminology at all, certain color examples were better remembered. She argues that these “natural” categories are generally favored in human learning and cognition. This work is often taken as support for universals of color perception, though since the Dani had no linguistic categories to sway them away from biologically primary colors, this cannot be taken as evidence against a potential linguistic influence on color perception.

The effects of language on color categorization could be seen in Kay and Kempton ( 1984 ), but any effects of language-specific color terms only surfaced under specific conditions, and the effects were not as robust as earlier researchers had hoped. Various proposals have been made to revise the Berlin and Kay approach in ways which accommodate linguistic relativity effects within a basically universally constrained system. Most notable of these is Vantage Theory, which seeks to explain multiple points of view—even within the putative universals of color perception—and how points of view may be linguistically mediated; see especially MacLaury ( 1991 , 1995 , 2000 ).

Work by Davies and colleagues has also expanded upon the work of Kay and Kempton ( 1984 ) by examining a variety of linguistic systems for denoting colors. They then test participants from these speech communities using various categorization tasks. For Turkish, see Oezgen and Davies ( 1998 ); for Setswana, English, and Russian, see Davies ( 1998 ), Davies and Corbett ( 1997 ), and Davies et al. ( 1998 ); see also Corbett and Davies ( 1997 ) for a discussion of method in language sampling for color terminology.

Especially within anthropology, there has been concern about the fundamental adequacy of the empirical method followed by Berlin and Kay (and later modifications). Jameson and DʼAndrade ( 1997 ) address the adequacy of the theory of color perception inherent in the use of the Munsell color system. Lucy ( 1997b ) criticizes most work on color terminology as insufficiently descriptive of the actual linguistic properties of the color terms themselves: without an adequate investigation into these properties, it is unclear what the effects may be of forcing reference with these terms into the Munsell system. The worry is that the Munsell system will not only standardize the coding of the responses, but actually create standardized and unnatural responses rather than allowing the terms to refer to their actual reference.

For a survey of recent work exploring color naming and its relationship to nonlinguistic cognition, see Kay and Regier ( 2006 ).

5.2. Shape Classification

In determining whether or not the Navajo shape classification system influenced sorting behavior, Carroll and Casagrande ( 1958 ) attempted to balance cultural factors across samples by using English-speaking and Navajo-speaking ethnic Navajo children. As a control group, English-speaking, middle-class American children were used. The results from triad classification (by either shape/function or color) were largely consistent with the Navajo verb classification, in that the Navajo-speaking Navajo children demonstrated a greater preference for shape sorting than English-speaking Navajo children. Note, however, that English-speaking middle-class children also patterned like Navajo-speaking children, suggesting to Carroll and Casagrande that cultural factors beyond language play an important role in such classification.

Lucy and Gaskins ( 2001 ) also use triad-type methods to compare Yucatecan children and adults with English-speaking Americans. Again, a broad consistency with each language's classification system is found, but interestingly, this only becomes prominent after age nine (see section 5.6 )

5.3. Conditional Reasoning

With basic reasoning processes, variation is more likely to be viewed as directly advantageous or disadvantageous, that is, essentially correct or incorrect. Whether the hypothesized cause is linguistic or otherwise, in modern academia, the burden of proof appropriately falls most heavily on the researcher hoping to demonstrate any potential absence (or “deficiency”) within a particular community.

The work of Alfred Bloom and his many detractors falls fully into this predicament. Bloom ( 1981 ) proposed that Chinese (unlike English) lacks a specific counterfactual construction and that this has led to reduced ability to engage in counterfactual reasoning. The debate was carried across several volumes of Cognition : Au ( 1983 , 1984 ), Bloom ( 1984 ), Liu ( 1985 ), Takano ( 1989 ); making use of different samples, these studies did not generally replicate Bloom's findings. 7 Unfortunately, there has been a tendency to interpret the various results (or lack thereof) as disconfirming linguistic relativity more generally rather than demonstrating a failure of a particular experimental design. Takano used Japanese speakers, who like Chinese speakers, lack a dedicated counterfactual construction, but found that their reasoning patterned like English speakers. More recently, Lardiere ( 1992 ) investigated Arabic speakers. Arabic patterns like English in that there is an explicit counterfactual construction, yet the Arabic participants performed like Bloom's original Chinese subjects on counterfactual reasoning. From these studies, both Takano and Lardiere conclude that the principal effect on counterfactual reasoning is traceable not to linguistic habit, but to cultural practices of reasoning, testing conventions, and the like.

Another conclusion one might draw from these studies is that we cannot automatically assume that either linguistic or nonlinguistic habit will be discern-able from the presence or absence of specialized linguistic constructions. Obviously, those Chinese and Japanese speakers trained in formal counterfactual reasoning must have found some means of expression. Conversely, the Arabic speakers need not have used their counterfactual construction in ways analogous to the ways of formally educated English speakers.

5.4. Number

Cardinal numbers.

One clear way in which languages vary is in their cardinal number systems. In addition to the obvious lack of larger numbers in many languages (at least as native vocabulary), languages also vary in their organization of these numbers. Various languages partially use a base twenty counting system and other languages appear to have relics of base five systems. But even within primarily base ten systems, there is variation of consistency and expression.

Miura ( 1987 ) argues that the generally superior mathematical abilities of school children in or from some cultures (especially East Asian) result at least in part from the transparency and exception-free nature of the base ten numerals used for counting, which children generally control prior to beginning formal education—see also the follow-up cross-linguistic studies: Miura and Okamoto ( 1989 ), Miura et al. ( 1988 ), Miura et al. ( 1993 ), Miura et al. ( 1994 ), Miura et al. ( 1999 ).

Saxton and Towse ( 1998 ) provide a more cautious conclusion, suggesting that the influence of native language on the task of learning place values is less than argued for by Miura and colleagues. Many other differences in performance were found across groups which were better accounted for as resulting from general cultural attitudes toward education and so on, than as the result of the linguistic number system.

Grammatical Number

On a grammatical level, languages vary in terms of their grammatical encoding of the number of entities in an event or scene. While this topic has not been widely taken up, the work of Lucy ( 1992a ) is noteworthy for its extensive consideration of attention to number in Mayan and English speakers. An extensive typological discussion of grammatical number, though without focus on issues of linguistic relativity, is provided by Corbett ( 2000 ). Lastly, Hill and Hill ( 1998 ) discuss the effects of culture on language (rather than linguistic relativity) for number marking (plurals), and in particular the “anti-Whorfian effect” they find in Uto-Aztecan.

Reference Frames

Currently, the primary area of linguistic relativity research in spatial domains is with reference frames (however, there is also the important developmental work on topological relations by Choi and Bowerman 1991 , see below).

Reference frames are the psychological or linguistic representation of relationships between entities in space. They require fixed points of reference, such as the speaker, a landmark, or an established direction. Within linguistics, the typology of reference frames is complicated, but most accounts include something like an intrinsic reference frame (whereby an object is located only with respect to an immediate point, e.g., The ball is next to the chair ) and various flavors of reference frames which make use of additional orientation (e.g., The ball is to my right of the chair or The ball is to the north of the chair ). Languages vary in terms of their habitually selected reference frames, and following the linguistic relativity hypothesis, speakers should also vary in their encoding spatial memories, making locational calculations, and so forth. For extensive work measuring event-related potential data (recordings at the scalp of electrical charges from brain activity during specific tasks), see the work of Taylor and colleagues: Taylor et al. ( 1999 ) and Taylor et al. ( 2001 ). These works compare the viewer/speaker-relative (or egocentric ) reference frame with the intrinsic.

Of note for being broadly comparative across diverse linguistic and cultural communities is the work reported in Pederson et al. ( 1998 ), which found correlations between habitual linguistic selection of reference frames and cognitive performance on spatial memory (and other) tasks. There were many studies within this same general project. Perhaps the most important to consult for the theoretical underpinnings for the project are Brown and Levinson ( 1993 ) and Levinson ( 1996 ). As pointed out by Li and Gleitman ( 2002 ), the populations reported as using an absolute/geo-cardinal ( north of …) reference frame were largely rural populations, and the populations using a speaker-relative/egocentric reference frame are largely urban, so there is a potential confound in the population samples between language and culture/environment. For a rebuttal to these concerns and Li and Gleitman's similar experiments, see Levinson et al. ( 2002 ); see also Pederson ( 1998 ) for a discussion of this urban/rural cultural split.

Motion Events

Talmy ( 1985 , 2000b ) identifies a typological contrast in the ways that languages encode basic motion events. To simplify, some languages such as the Romance languages commonly encode the fact of motion and the basic path with the main verb (e.g., to enter , to ascend , etc.). In contrast, Germanic and many other languages most commonly encode the fact of motion along with the manner of motion in the verb (e.g., to wiggle ), and the path is expressed elsewhere.

Slobin ( 1991 , 1996 ) considers the cognitive consequences of these linguistic patterns for English and Spanish speakers. Slobin ( 2000 ) extends this approach to French, Hebrew, Russian, and Turkish. Gennari et al. ( 2002 ) and Malt, Sloman, and Gennari ( 2003 ) examine these contrasts experimentally and argue for some effects of one's native language pattern on certain nonlinguistic tasks.

While spatial relationships have been extensively studied for linguistic relativity effects, the effects of different temporal encoding have received much less attention. In part, this may be attributed to the relative difficulty of developing research instruments. An obvious difference cross-linguistically is whether or not a language grammatically encodes tense. Bohnemeyer ( 1998 ) discusses the lack of tense-denoting constructions in Yucatec Mayan and contrasts this with German speakers observing the same video stimuli; nonetheless, both samples appeared to have encoded similar event orderings in memory. Languages also have some variation in preferred metaphors for talking about time. Boroditsky ( 2000 , 2001 ) argues that Mandarin Chinese speakers have a different metaphor for time (vertical) and this appears to influence their nonlinguistic encoding as well.

5.7. Developmental Studies

Ultimately, any linguistic relativity effects must be explained in terms of the acquisition of linguistic categories and the effects on cognitive development.

Choi and Bowerman ( 1991 ) and Bowerman and Choi ( 2001 ) contrast early lexical acquisition of Korean and English spatial terms, principally those expressing contact, closure, and similar concepts. Korean-speaking adults use spatial terms to categorize subtypes of these different relationships in very different ways from English-speaking adults. Perhaps surprisingly, Choi and Bowerman report that Korean-speaking children as young as two demonstrate linguistic patterning more like the Korean-speaking adults than like the English-speaking children (and vice versa). This suggests that even in fairly early lexical acquisition, children show remarkable sensitivity to the specific language input rather than relying on purportedly universal cognitive categorizations and fitting the language categories onto these.

Lowenstein and Gentner ( 1998 ), Gentner and Loewenstein ( 2002 ), and Gentner and Boroditsky ( 2001 ) argue that metaphor and analogical reasoning are key parts of concept development and early word meaning. To the extent that these are cross-linguistically variable, it can be argued that linguistic relativity effects may be present especially for abstract reasoning which most depends on relational terminology and analogy.

As mentioned in the section on shape classification, Lucy and Gaskins ( 2001 ) look at the age of development of language-particular patterns in shape versus material sorting tasks. Assuming one can extrapolate from their data, the critical age at which language helps to direct nonlinguistic behavior (for these sorts of tasks) is around ages 7–9. This suggests that the acquisition of language categories need not immediately manifest cognitive effects in nonlinguistic domains, but rather that there maybe a period in which the linguistic categories are initially more solely linguistic and then eventually the analogy from language to other types of categorization is drawn. It may also reflect a greater dependence on linguistically mediated internal thought, à la Vygotsky.

Susan Goldin-Meadow and colleagues have examined the interplay of gesture, home sign, and conventional language use and their relationships to underlying (and developing) cognitive representations. A good recent summary may be found in Goldin-Meadow ( 2002 ) and the references within. Zheng and Goldin-Meadow ( 2002 ) examine the similarities across cultures in home sign despite notable differences in the adult spoken languages. These commonalities suggest what the underlying conceptual categories may be in children prior to acquiring the “filter” provided by the model of a specific language.

Working with English-speaking children and language acquisition delayed deaf children, de Villiers and de Villiers ( 2000 ) argue that language has a vital role in the development of understandings of false beliefs—at least insofar as demonstrated in unseen displacement. (For example, the puppet doesn't see that I replaced the crayons in the crayon box with a key; what does the puppet think is in the crayon box?) Language is eminently suited for the representation of counterfactual and alternative beliefs, so it is unclear whether it is the specifics of language acquisition or just general exposure to alternatives that happen to come through the medium of language which might be driving this development. For a summary of the work by Gopnik and colleagues on the potential interactions of language and cognitive development, especially around ages 1–2, see Gopnik ( 2001 ).

5.8. Sign Language versus Spoken Language

Lastly, what of the medium of the language itself? Might the mechanical constraints of spoken language versus sign language have their own influences? Working with native ASL signers and English speakers on mental rotation tasks, Emmorey, Klima, and Hickok ( 1998 ) show evidence that the vast experience of signers in understanding their interlocutors' spatial perspective during signing has given them some advantage in nonlinguistic rotation tasks compared with nonsigners.

6. Future Directions

As can be seen from the above discussion, the issue of linguistic relativity is as open a question as it is broad. However, as empirically driven models of human cognition become increasingly detailed, work within linguistic relativity (and Cognitive Linguistics generally) becomes increasingly specific in its description of cognitive mechanisms.

The question “Does language influence thought?” is being replaced by a battery of questions about whether a given feature of a specific language influences particular cognitive operations, what the exact cognitive mechanisms are which give rise to this influence, and how we can most precisely characterize the nature of this influence? Rather than this being a step away from the “big picture” of human cognition, this general trend toward increasingly precise definitions and, ideally, more falsifiable hypotheses leads us to a simply more reliable understanding of cognition and the role of language within it.

As we discover more of the specific interactions between language and the rest of the cognitive systems, there is a need to understand the time course of this development. Except for Lucy and Gaskins ( 2001 ) and some of the home sign studies, there has been virtually no attempt to determine the time course of any linguistic relativity effects. If language influences a particular cognitive operation or conceptualization, does it do so upon acquisition of the language model, shortly subsequent to this acquisition, or is there a gradual “internalization” (in Vygotskian terms) of the linguistic structure as something more than a learned code?

One must also wonder whether certain linguistic construals more readily have influences beyond language than others. For example, is spatial categorization more likely to be influenced by language than color categorization is, or vice versa? If some domains are more linguistically sensitive, what do these domains have in common?

These are all broad questions and are unlikely to be resolved in the immediate future. However, as research in linguistic relativity becomes increasingly mainstream within psychology and linguistics, it seems certain that we will understand ever more of the complexities between language and thought.

Many more recent writings by Alford on Whorf, linguistic relativity, and related topics can be found on Alford's Web site: http://www.enformy.com/alford.htm.

This idea was apparently insufficiently discredited as it has more recently resurfaced in the popular press with Shlain ( 1998 )—where it is now associated with the demise of polytheism and the claimed consequent surge of misogyny in European history.

Anecdotally, I can report that subjects in spatial reference frame experiments would use their linguistically dominant frame of reference in nonlinguistic tasks but would switch when they heard an alternate frame of reference used immediately before the task. (Specifically, when an assistant erroneously used nonneutral language in an example.) In subsequent tasks, with no reference frame language repeated, the subjects could switch over to what might well have been a more default reference frame for such tasks. Of course, these subject results are not coded with other subjects, and this dictated extreme care in controlling the immediately preceding linguistic environment during experimental sessions.

College students (especially those participating for credit in an introductory psychology class!) are infamous for trying to second guess the “hidden” purpose of an experiment. Surely, such subjects are less directly comparable with the perhaps experimentally less savvy subjects drawn from other populations.

Li and Gleitman ( 2002 ) changed “small procedural details” (see their footnote 5) in this experiment—notably they eliminated the distance between the tables—and report different results. Although they do not attribute the different results to these changes, but rather to other uncontrolled variables in the original study, the control of the experimental setup clearly can be critical for evaluating the results.

The linguistic parallels with basic operations in visual perception imply a bias favoring the building of linguistic categories from more fundamental cognitive categories rather than any particular influence from language to cognition.

Cara and Politzer ( 1993 ) also found no correspondence of language to reasoning with Chinese and English speakers on counterfactual reasoning tasks, though the design seems uninfluenced by the debate in Cognition .

Alford, Dan K. H. 1978 . The demise of the Whorf hypothesis.   Berkeley Linguistics Society 14: 485–99.

Au, Terry Kit-Fong. 1983 . Chinese and English counterfactuals: The Sapir-Whorf hypothesis revisited.   Cognition 15: 155–87.

Au, Terry Kit-Fong. 1984 . Counterfactuals: In reply to Alfred Bloom.   Cognition 17: 289–302.

Berlin, Brent. 1978 . Ethnobiological classification. In Eleanor Rosch and Barbara B. Lloyd, eds., Cognition and categorization 9–26. Hillsdale, NJ: Lawrence Erlbaum.

Google Scholar

Google Preview

Berlin, Brent, Dennis E. Breedlove, and Peter H. Raven. 1973 . General principles of classification and nomenclature in folk biology.   American Anthropologist 75: 214–42.

Berlin, Brent, and Paul Kay. 1969 . Basic color terms: Their universality and evolution . Berkeley: University of California Press.

Bickel, Balthasar. 2000 . Grammar and social practice: On the role of ‘culture’ in linguistic relativity. In Susanne Niemeier and René Dirven, eds., Evidence for linguistic relativity 161–91. Amsterdam: John Benjamins.

Bloom, Alfred H. 1981 . The linguistic shaping of thought: A study in the impact of language on thinking in China and the West . Hillsdale, NJ: Lawrence Erlbaum.

Bloom, Alfred H. 1984 . Caution—the words you use may affect what you say: A response to Au.   Cognition 17: 275–87.

Blount, Ben G. 1993 . Cultural bases of folk classification systems. In Jeanette Altarriba, ed., Cognition and culture: A cross-cultural approach to psychology 3–22. Amsterdam: Elsevier.

Bohnemeyer, Jürgen. 1998. Time relations in discourse: Evidence from a comparative approach to Yukatek Maya. PhD dissertation, Katholieke Universiteit Brabant, Netherlands.

Boroditsky, Lera. 2000 . Metaphoric structuring: Understanding time through spatial metaphors.   Cognition 75: 1–28.

Boroditsky, Lera. 2001 . Does language shape thought? Mandarin and English speakers' conceptions of time.   Cognitive Psychology 43: 1–22.

Bowerman, Melissa , and Soonja Choi. 2001 . Shaping meanings for language: Universal and language-specific in the acquisition of spatial semantic categories. In Melissa Bowerman and Stephen C. Levinson, eds., Language acquisition and conceptual development 475–511. Cambridge: Cambridge University Press.

Brown, Penelope, and Stephen C. Levinson. 1993 . “Uphill” and “downhill” in Tzeltal.   Journal of Linguistic Anthropology 3: 46–74.

Brown, Roger Langham. 1967 . Wilhelm von Humboldt's conception of linguistic relativity: Janua linguarum . The Hague: Mouton.

Brown, Roger W., and Eric. H. Lenneberg. 1958 . Studies in linguistic relativity. In Eleanor E. Maccoby, Theodore M. Newcomb, and Eugene L. Hartley, eds., Readings in social psychology 9–18. New York: Holt, Rinehart and Winston.

Brugman, Claudia, and Monica Macaulay. 1986 . Interacting semantic systems: Mixtec expressions of location.   Berkeley Linguistics Society 12: 315–27.

Cara, Francesco, and Guy Politzer. 1993 . A comparison of conditional reasoning in English and Chinese. In Jeanette Altarriba, ed., Cognition and culture: A cross-cultural approach to psychology 283–97. Amsterdam: Elsevier.

Carroll, John B., and Joseph B. Casagrande. 1958 . The function of language classifications in behavior. In Eleanor. E. Maccoby, Theodore M. Newcomb, and Eugene L. Hartley, eds., Readings in social psychology 18–31. New York: Holt, Rinehart and Winston.

Carter, Anthony T. 1973 . A comparative analysis of systems of kinship and marriage in South Asia.   Proceedings of the Royal Anthropological Institute of Great Britain and Ireland : 29–54.

Chafe, Wallace. 2000 . Loci of diversity and convergence in thought and language. In Martin Pütz and Marjolijn Verspoor, eds., Explorations in linguistic relativity 101–23. Amsterdam: John Benjamins.

Choi, Soonja, and Melissa Bowerman. 1991 . Learning to express motion events in English and Korean: The influence of language-specific lexicalization patterns.   Cognition 41: 83–121.

Corbett, Greville G. 2000 . Number . Cambridge: Cambridge University Press.

Corbett, Greville G., and Ian R. L. Davies. 1997 . Establishing basic color terms: Measures and techniques. In C. L. Hardin and Luisa Maffi, eds., Color categories in thought and language 197–223. Cambridge: Cambridge University Press.

Danziger, Eve, and Eric Pederson. 1998 . Through the looking glass: Literacy, writing systems and mirror image discrimination.   Written Language and Literacy 1: 153–64.

Davies, Ian R. L. 1998 . A study of colour grouping in three languages: A test of the linguistic relativity hypothesis.   British Journal of Psychology 89: 433–52.

Davies, Ian R. L., and Greville G. Corbett. 1997 . A cross-cultural study of colour grouping: Evidence for weak linguistic relativity.   British Journal of Psychology 88: 493–517.

Davies, Ian R. L., Paul T. Sowden, David T. Jerrett, Tiny Jerrett, and Greville G. Corbett. 1998 . A cross-cultural study of English and Setswana speakers on a colour triads task: A test of the Sapir-Whorf hypothesis.   British Journal of Psychology 89: 1–15.

de Villiers, Jill G., and Peter A. de Villiers. 2000 . Linguistic determinism and the understanding of false beliefs. In Peter Mitchell and Kevin John Riggs, eds., Children's reasoning and the mind 191–228. East Sussex: Psychology Press.

Duranti, Alessandro. 1997 . Linguistic anthropology . Cambridge: Cambridge University Press.

Emmorey, Karen, Edward Klima, and Gregory Hickok. 1998 . Mental rotation within linguistic and non-linguistic domains in users of American sign language.   Cognition 68: 221–46.

Enfield, Nick J. 2000 . On linguocentrism. In Martin Pütz and Marjolijn Verspoor, eds., Explorations in linguistic relativity 125–57. Amsterdam: John Benjamins.

Foley, William A. 1997 . Anthropological linguistics: An introduction . Oxford: Basil Blackwell.

Geeraerts, Dirk, Stefan Grondelaers, and Peter Bakema. 1994 . The structure of lexical variation: Meaning, naming, and context . Berlin: Mouton de Gruyter.

Gennari, Silvia P., Steven A. Sloman, Barbara C. Malt, and W. Tecumseh Fitch. 2002 . Motion events in language and cognition.   Cognition 83: 49–79.

Gentner, Dedre, and Lera Boroditsky. 2001 . Individuation, relativity, and early word learning. In Melissa Bowerman and Stephen C. Levinson, eds., Language acquisition and conceptual development 215–56. Cambridge: Cambridge University Press.

Gentner, Dedre, and Jeffrey Loewenstein. 2002 . Relational language and relational thought: Language, literacy, and cognitive development. In Eric Amsel and James P. Byrnes, eds., The development and consequences of symbolic communication 87–120. Mahwah, NJ: Lawrence Erlbaum.

Goldin-Meadow, Susan. 2002 . From thought to hand: Structured and unstructured communication outside of conventional language. In Eric Amsel and James P. Byrnes, eds., Language, literacy, and cognitive development: The development and consequences of symbolic communication 121–50. Mahwah, NJ: Lawrence Erlbaum.

Goody, Jack, and Ian Watt. 1962 . The consequences of literacy.   Comparative Studies in Sociology and History 5: 304–45.

Gopnik, Alison. 2001 . Theories, language and culture: Whorf without wincing. In Melissa Bowerman and Stephen C. Levinson, eds., Language acquisition and conceptual development 45–69. Cambridge: Cambridge University Press.

Graff, Harvey J. 1987 . The legacies of literacy . Bloomington: Indiana University Press.

Hanks, William F. 1990 . Referential practice: Language and lived space among the Maya . Chicago: University of Chicago Press.

Hardin, C. L., and Luisa Maffi. 1997 . Color categories in thought and language . Cambridge: Cambridge University Press.

Heider, Eleanor R. 1971 . “Focal” color areas and the development of color names.   Developmental Psychology 4: 447–55.

Heider, Eleanor R. 1972 . Universals in color naming and memory.   Journal of Experimental Psychology 93: 10–20.

Heider, Eleanor R., and Donald C. Olivier. 1972 . The structure of the color space in naming and memory in two languages.   Cognitive Psychology 3: 337–54.

Hering, Ewald. 1964 . Outlines of a theory of the light sense . Cambridge, MA: Harvard University Press.

Hill, Jane H., and Kenneth C. Hill. 1998 . Culture influencing language: Plurals of Hopi kin terms in comparative Uto-Aztecan perspective.   Journal of Linguistic Anthropology 7: 166–80.

Hill, Jane H., and Bruce Mannheim. 1992 . Language and world view.   Annual Review of Anthropology 21: 381–406.

House, Juliane. 2000 . Linguistic relativity and translation. In Martin Pütz and Marjolijn Verspoor, eds., Explorations in linguistic relativity 69–88. Amsterdam: John Benjamins.

Hunn, Eugene S. 1977 . Tzeltal folk zoology: The classification of discontinuities in nature . New York: Academic Press.

Hunn, Eugene S. 1982 . The utilitarian factor in folk biological classification.   American Anthropologist 84: 830–47.

Hunt, Earl, and Franca Agnoli. 1991 . The Whorfian hypothesis: A cognitive psychology perspective.   Psychological Review 98: 377–89.

Jameson, Kimberly, and Roy G. DʼAndrade. 1997 . It's not really red, green, yellow, blue: An inquiry into perceptual color space. In C. L. Hardin and Luisa Maffi, eds., Color categories in thought and language 295–319. Cambridge: Cambridge University Press.

Kay, Paul. 1996 . Intra-speaker relativity. In John J. Gumperz and Stephen C. Levinson, eds., Rethinking linguistic relativity 97–114. Cambridge: Cambridge University Press.

Kay, Paul, and Willett Kempton. 1984 . What is the Sapir-Whorf hypothesis?   American Anthropologist 86: 65–79.

Kay, Paul, and Chad K. McDaniel. 1978 . The linguistic significance of the meanings of basic color terms.   Language 54: 610–46.

Kay, Paul, and Terry Regier. 2006 . Language, thought and color: Recent developments.   Trends in Cognitive Sciences 10: 51–54.

Koerner, Konrad E. F. 2000 . Towards a ‘full pedigree’ of the ‘Sapir-Whorf Hypothesis’: From Locke to Lucy. In Martin Pütz and Marjolijn Verspoor, eds., Explorations in linguistic relativity 1–23. Amsterdam: John Benjamins.

Lakoff, George. 1987 . Women, fire, and dangerous things: What categories reveal about the mind . Chicago: University of Chicago Press.

Langacker, Ronald W. 1987 . Nouns and verbs.   Language 63: 53–94.

Langacker, Ronald W. 1991 . Foundations of cognitive grammar . Vol. 2, Descriptive application . Stanford, CA: Stanford University Press.

Lardiere, Donna. 1992 . On the linguistic shaping of thought: Another response to Alfred Bloom.   Language in Society 21: 231–51.

Lee, Penny. 1996 . The Whorf theory complex: A critical reconstruction . Amsterdam: John Benjamins.

Lee, Penny. 2000 . When is ‘linguistic relativity’ Whorf's linguistic relativity? In Martin Pütz and Marjolijn Verspoor, eds., Explorations in linguistic relativity 45–68. Amsterdam: John Benjamins.

Lenneberg, Eric H., and John M. Roberts. 1956 . The language of experience: A study in methodology . International Journal of American Linguistics Memoir, no. 13. Baltimore, MD: Waverly Press.

Levinson, Stephen C. 1996 . Frames of references and Molyneux's question: Crosslinguistic evidence. In Paul Bloom, Mary A. Peterson, Lynn Nadel, and Merrill Garrett, eds., Language and space 109–69. Cambridge, MA: MIT Press.

Levinson, Stephen C., Sotaro Kita, Daniel B. M. Hauna, and Björn H. Rasch. 2002 . Returning the tables: Language affects spatial reasoning.   Cognition 84: 155–88.

Li, Peggy, and Lila Gleitman. 2002 . Turning the tables: Language and spatial reasoning.   Cognition 83: 265–94.

Liu, Lisa Gabern. 1985 . Reasoning counterfactually in Chinese: Are there any obstacles?   Cognition 21: 239–70.

Loftus, Elizabeth F. 1975 . Leading questions and the eyewitness report.   Cognitive Psychology 7: 560–72.

Losonsky, Michael, ed. 1999 . Wilhelm von Humboldt: On language: On the diversity of human language construction and its influence on the mental development of the human species . Cambridge: Cambridge University Press.

Lowenstein, Jeffrey, and Dedre Gentner. 1998 . Relational language facilitates analogy in children. Proceedings of the Twentieth Annual Conference of the Cognitive Science Society 615–20. Mahwah, NJ.: Lawrence Erlbaum.

Lucy, John A. 1992 a. Grammatical categories and cognition: A case study of the linguistic relativity hypothesis . Cambridge: Cambridge University Press.

Lucy, John A. 1992 b. Language diversity and thought: A reformulation of the linguistic relativity hypothesis . Cambridge: Cambridge University Press.

Lucy, John A. 1997 a. Linguistic relativity.   Annual Review of Anthropology 26: 291–312.

Lucy, John A. 1997 b. The linguistics of “color.” In C. L. Hardin and Luisa Maffi, eds., Color categories in thought and language 320–46. Cambridge: Cambridge University Press.

Lucy, John A., and Suzanne Gaskins. 2001 . Grammatical categories and the development of classification preferences: A comparative approach. In Melissa Bowerman and Stephen C. Levinson, eds., Language acquisition and conceptual development 257–83. Cambridge: Cambridge University Press.

Lucy, John A., and Richard A. Shwedder. 1988 . The effect of incidental conversation on memory for focal colors.   American Anthropologist 90: 923–31.

Luria, Aleksandr Romanovich. 1976 . Cognitive development, its cultural and social foundations . Cambridge, MA: Harvard University Press.

MacLaury, Robert E. 1991 . Exotic color categories: Linguistic relativity to what extent?   Journal of Linguistic Anthropology 1: 26–51.

MacLaury, Robert E. 1995 . Vantage theory. In John R. Taylor and Robert E. MacLaury, eds., Language and the cognitive construal of the world 231–76. Berlin: Mouton de Gruyter.

MacLaury, Robert E. 2000 . Linguistic relativity and the plasticity of categorization: Universalism in a new key. In Martin Pütz and Marjolijn H. Verspoor, eds., Explorations in linguistic relativity 251–93. Amsterdam: John Benjamins.

Malt, Barbara C., Steven A. Sloman, and Silvia P. Gennari. 2003 . Speaking versus thinking about objects and actions. In Dedre Gentner and Susan Goldin-Meadow, eds., Language in mind: Advances in the study of language and thought 81–111. Cambridge, MA: MIT Press.

Manchester, Martin L. 1985 . The philosophical foundations of Humboldt's linguistic doctrines . Amsterdam: John Benjamins.

Miura, Irene T. 1987 . Mathematics achievement as a function of language.   Journal of Educational Psychology 79: 79–82.

Miura, Irene T., Chungsoon C. Kim, Chih-mei Chang, and Yukari Okamoto. 1988 . Effects of language characteristics on children's cognitive representation of number: Cross-national comparisons.   Child Development 59: 1445–50.

Miura, Irene T., and Yukari Okamoto. 1989 . Comparisons of U.S. and Japanese first graders' cognitive representation of number and understanding of place value.   Journal of Educational Psychology 81: 109–14.

Miura, Irene T., Yukari Okamoto, Chungsoon C. Kim, Chih-Mei Chang, Marcia Steere, and Michel Fayol. 1994 . Comparisons of children's cognitive representation of number: China, France, Japan, Korea, Sweden, and the United States.   International Journal of Behavioral Development 17: 401–11.

Miura, Irene T., Yukari Okamoto, Chungsoon C. Kim, Marcia Steere, and Michel Fayol. 1993 . First graders' cognitive representation of number and understanding of place value: Cross-national comparisons: France, Japan, Korea, Sweden, and the United States.   Journal of Educational Psychology 85: 24–30.

Miura, Irene T., Yukari Okamoto, Vesna Vlahovic-Stetic, Chungsoon C. Kim, and Jong Hye Han. 1999 . Language supports for children's understanding of numerical fractions: Cross-national comparisons.   Journal of Experimental Child Psychology 74: 356–65.

Morais, José, Luz Cary, Jésus Alegria, and Paul Bertelson. 1979 . Does awareness of speech as a sequence of phones arise spontaneously?   Cognition 7: 323–31.

Oezgen, Emre, and Ian R. L. Davies. 1998 . Turkish color terms: Tests of Berlin and Kay's theory of color universals and linguistic relativity.   Linguistics 36: 919–56.

Olson, David R. 1991 . Literacy as metalinguistic activity. In David R. Olson and Nancy Torrance, eds., Literacy and orality 251–70. Cambridge: Cambridge University Press.

Olson, David R. 2002 . What writing does to the mind. In Eric Amsel and James P. Byrnes, eds., Language, literacy, and cognitive development: The development and consequences of symbolic communication 153–65. Mahwah, NJ: Lawrence Erlbaum.

Ong, Walter J. 1992 . Writing is a technology that restructures thought. In Pam Downing, Susan D. Lima, and Michael Noonan, eds., The linguistics of literacy 293–319. Amsterdam: John Benjamins.

Pederson, Eric. 1993 . Geographic and manipulable space in two Tamil linguistic systems. In Andrew U. Frank and Irene Campari, eds., Spatial information theory 294–311. Berlin: Springer Verlag.

Pederson, Eric. 1995 . Language as context, language as means: Spatial cognition and habitual language use.   Cognitive Linguistics 6: 33–62.

Pederson, Eric. 1998 . Spatial language, reasoning, and variation across Tamil communities. In Petr Zima and Vladimír Tax, eds., Language and location in space and time 111–19. Munich: Lincom Europa.

Pederson, Eric. 2003 . Mirror-image discrimination among nonliterate, monoliterate, and biliterate Tamil speakers.   Written Language and Literacy 6: 71–91.

Pederson, Eric, Eve Danziger, David Wilkins, Stephen Levinson, Sotaro Kita, and Gunter Senft. 1998 . Semantic typology and spatial conceptualization.   Language 74: 557–89.

Read, Charles, Yun-Fei Zhang, Hong-Yin Nie, and Bao-Qing Ding. 1986 . The ability to manipulate speech sounds depends on knowing alphabetic writing.   Cognition 24: 31–44.

Rosch, Eleanor. 1973 . Natural categories.   Cognitive Psychology 4: 328–50.

Saxton, Matthew, and John N. Towse. 1998 . Linguistic relativity: The case of place value in multi-digit numbers.   Journal of Experimental Child Psychology 69: 66–79.

Scinto, Leonard F. M. 1986 . Written language and psychological development . New York: Academic Press.

Scribner, Sylvia, and Michael Cole. 1981 . The psychology of literacy . Cambridge, MA: Harvard University Press.

Shlain, Leonard. 1998 . The alphabet versus the goddess: The conflict between word and image . New York: Viking.

Silverstein, Michael. 1985 . Language and the culture of gender: At the intersection of structure, usage, and ideology. In Elizabeth Mertz and Richard J. Parmentier, eds., Semiotic mediation: Sociocultural and psychological perspectives 219–59. Orlando, FL: Academic Press.

Silverstein, Michael. 1987 . Cognitive implications of a referential hierarchy. In Maya Hickmann, ed., Social and functional approaches to language 125–64. Orlando, FL: Academic Press.

Slobin, Dan I. 1991 . Learning to think for speaking: Native language, cognition, and rhetorical style.   Pragmatics 1: 7–25.

Slobin, Dan I. 1996 . Two ways to travel: Verbs of motion in English and Spanish. In Masayoshi Shibatani and Sandra A. Thompson, eds., Grammatical constructions: Their form and meaning 195–219. Oxford: Oxford University Press.

Slobin, Dan I. 2000 . Verbalized events: A dynamic approach to linguistic relativity and determinism. In Susanne Niemeier and René Dirven, eds., Evidence for linguistic relativity 107–38. Amsterdam: John Benjamins.

Smith, Marion V. 1996 . Linguistic relativity: On hypotheses and confusions.   Communication & Cognition 29: 65–90.

Takano, Yohtaro. 1989 . Methodological problems in cross-cultural studies of linguistic relativity.   Cognition 31: 141–62.

Talmy, Leonard. 1977 . Rubber-sheet cognition in language.   Chicago Linguistic Society 13: 612–28.

Talmy, Leonard. 1978 . The relation of grammar to cognition—a synopsis. In David Waltz ed., Proceedings of TINLAP -2 14–24. New York: Association for Computing Machinery.

Talmy, Leonard. 1985 . Lexicalization patterns: Semantic structure in lexical form. In Timothy Shopen, ed., Language typology and syntactic description , vol. 3, Grammatical categories and the lexicon 57–149. Cambridge: Cambridge University Press.

Talmy, Leonard. 2000 a. Toward a cognitive semantics . Vol. 1, Concept structuring systems . Cambridge, MA: MIT Press.

Talmy, Leonard. 2000 b. Toward a cognitive semantics . Vol. 2, Typology and process in concept structuring . Cambridge, MA: MIT Press.

Taylor, Holly A., Robert R. Faust, Tatiana Sitnikova, Susan J. Naylor, and Phillip J. Hol-comb. 2001 . Is the donut in front of the car? An electrophysiological study examining spatial reference frame processing.   Canadian Journal of Experimental Psychology 55: 175–84.

Taylor, Holly A., Susan J. Naylor, Robert R. Faust, and Phillip J. Holcomb. 1999 . “Could you hand me those keys on the right?” Disentangling spatial reference frames using different methodologies.   Spatial Cognition and Computation 1: 381–97.

Whorf, Benjamin L. 1956 . Language, thought, and reality: Selected writings of Benjamin Lee Whorf . Cambridge, MA: MIT Press.

Whorf, Benjamin L., and George L. Trager. [1938] 1996 . Report on linguistic research in the department of Anthropology of Yale University for the term Sept. 1937–June 1938. In Penny Lee, The Whorf theory complex: A critical reconstruction 251–80. Amsterdam: John Benjamins.

Zheng, Mingyu, and Susan Goldin-Meadow. 2002 . Thought before language: How deaf and hearing children express motion events across cultures.   Cognition 85: 145–74.

  • About Oxford Academic
  • Publish journals with us
  • University press partners
  • What we publish
  • New features  
  • Open access
  • Institutional account management
  • Rights and permissions
  • Get help with access
  • Accessibility
  • Advertising
  • Media enquiries
  • Oxford University Press
  • Oxford Languages
  • University of Oxford

Oxford University Press is a department of the University of Oxford. It furthers the University's objective of excellence in research, scholarship, and education by publishing worldwide

  • Copyright © 2024 Oxford University Press
  • Cookie settings
  • Cookie policy
  • Privacy policy
  • Legal notice

This Feature Is Available To Subscribers Only

Sign In or Create an Account

This PDF is available to Subscribers Only

For full access to this pdf, sign in to an existing account, or purchase an annual subscription.

Library homepage

  • school Campus Bookshelves
  • menu_book Bookshelves
  • perm_media Learning Objects
  • login Login
  • how_to_reg Request Instructor Account
  • hub Instructor Commons

Margin Size

  • Download Page (PDF)
  • Download Full Book (PDF)
  • Periodic Table
  • Physics Constants
  • Scientific Calculator
  • Reference & Cite
  • Tools expand_more
  • Readability

selected template will load here

This action is not available.

Social Sci LibreTexts

3.1: Linguistic Relativity- The Sapir-Whorf Hypothesis

  • Last updated
  • Save as PDF
  • Page ID 75159

  • Manon Allard-Kropp
  • University of Missouri–St. Louis

\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

\( \newcommand{\Span}{\mathrm{span}}\)

\( \newcommand{\id}{\mathrm{id}}\)

\( \newcommand{\kernel}{\mathrm{null}\,}\)

\( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\)

\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\)

\( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

\( \newcommand{\vectorA}[1]{\vec{#1}}      % arrow\)

\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}}      % arrow\)

\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vectorC}[1]{\textbf{#1}} \)

\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

Learning Objectives

After completing this module, students will be able to:

1. Define the concept of linguistic relativity

2. Differentiate linguistic relativity and linguistic determinism

3. Define the Sapir-Whorf Hypothesis (against more pop-culture takes on it) and situate it in a broader theoretical context/history

4. Provide examples of linguistic relativity through examples related to time, space, metaphors, etc.

In this part, we will look at language(s) and worldviews at the intersection of language & thoughts and language & cognition (i.e., the mental system with which we process the world around us, and with which we learn to function and make sense of it). Our main question, which we will not entirely answer but which we will examine in depth, is a chicken and egg one: does thought determine language, or does language inform thought?

We will talk about the Sapir-Whorf Hypothesis; look at examples that support the notion of linguistic relativity (pronouns, kinship terms, grammatical tenses, and what they tell us about culture and worldview); and then we will more specifically look into how metaphors are a structural component of worldview, if not cognition itself; and we will wrap up with memes. (Can we analyze memes through an ethnolinguistic, relativist lens? We will try!)

3.1 Linguistic Relativity: The Sapir-Whorf Hypothesis

In the 1920s, Benjamin Whorf was a graduate student studying with linguist Edward Sapir at Yale University in New Haven, Connecticut. Sapir, considered the father of American linguistic anthropology, was responsible for documenting and recording the languages and cultures of many Native American tribes, which were disappearing at an alarming rate. This was due primarily to the deliberate efforts of the United States government to force Native Americans to assimilate into the Euro-American culture. Sapir and his predecessors were well aware of the close relationship between culture and language because each culture is reflected in and influences its language. Anthropologists need to learn the language of the culture they are studying in order to understand the world view of its speakers. Whorf believed that the reverse is also true, that a language affects culture as well, by actually influencing how its speakers think. His hypothesis proposes that the words and the structures of a language influence how its speakers think about the world, how they behave, and ultimately the culture itself. (See our definition of culture in Part 1 of this document.) Simply stated, Whorf believed that human beings see the world the way they do because the specific languages they speak influence them to do so.

He developed this idea through both his work with Sapir and his work as a chemical engineer for the Hartford Insurance Company investigating the causes of fires. One of his cases while working for the insurance company was a fire at a business where there were a number of gasoline drums. Those that contained gasoline were surrounded by signs warning employees to be cautious around them and to avoid smoking near them. The workers were always careful around those drums. On the other hand, empty gasoline drums were stored in another area, but employees were more careless there. Someone tossed a cigarette or lighted match into one of the “empty” drums, it went up in flames, and started a fire that burned the business to the ground. Whorf theorized that the meaning of the word empty implied to the worker that “nothing” was there to be cautious about so the worker behaved accordingly. Unfortunately, an “empty” gasoline drum may still contain fumes, which are more flammable than the liquid itself.

Whorf ’s studies at Yale involved working with Native American languages, including Hopi. The Hopi language is quite different from English, in many ways. For example, let’s look at how the Hopi language deals with time. Western languages (and cultures) view time as a flowing river in which we are being carried continuously away from a past, through the present, and into a future. Our verb systems reflect that concept with specific tenses for past, present, and future. We think of this concept of time as universal, that all humans see it the same way. A Hopi speaker has very different ideas and the structure of their language both reflects and shapes the way they think about time. The Hopi language has no present, past, or future tense. Instead, it divides the world into what Whorf called the manifested and unmanifest domains. The manifested domain deals with the physical universe, including the present, the immediate past and future; the verb system uses the same basic structure for all of them. The unmanifest domain involves the remote past and the future, as well as the world of desires, thought, and life forces. The set of verb forms dealing with this domain are consistent for all of these areas, and are different from the manifested ones. Also, there are no words for hours, minutes, or days of the week. Native Hopi speakers often had great difficulty adapting to life in the English speaking world when it came to being “on time” for work or other events. It is simply not how they had been conditioned to behave with respect to time in their Hopi world, which followed the phases of the moon and the movements of the sun.

In a book about the Abenaki who lived in Vermont in the mid-1800s, Trudy Ann Parker described their concept of time, which very much resembled that of the Hopi and many of the other Native American tribes. “They called one full day a sleep, and a year was called a winter. Each month was referred to as a moon and always began with a new moon. An Indian day wasn’t divided into minutes or hours. It had four time periods—sunrise, noon, sunset, and midnight. Each season was determined by the budding or leafing of plants, the spawning of fish, or the rutting time for animals. Most Indians thought the white race had been running around like scared rabbits ever since the invention of the clock.”

The lexicon , or vocabulary, of a language is an inventory of the items a culture talks about and has categorized in order to make sense of the world and deal with it effectively. For example, modern life is dictated for many by the need to travel by some kind of vehicle—cars, trucks, SUVs, trains, buses, etc. We therefore have thousands of words to talk about them, including types of vehicles, models, brands, or parts.

The most important aspects of each culture are similarly reflected in the lexicon of its language. Among the societies living in the islands of Oceania in the Pacific, fish have great economic and cultural importance. This is reflected in the rich vocabulary that describes all aspects of the fish and the environments that islanders depend on for survival. For example, in Palau there are about 1,000 fish species and Palauan fishermen knew, long before biologists existed, details about the anatomy, behavior, growth patterns, and habitat of most of them—in many cases far more than modern biologists know even today. Much of fish behavior is related to the tides and the phases of the moon. Throughout Oceania, the names given to certain days of the lunar months reflect the likelihood of successful fishing. For example, in the Caroline Islands, the name for the night before the new moon is otolol , which means “to swarm.” The name indicates that the best fishing days cluster around the new moon. In Hawai`i and Tahiti two sets of days have names containing the particle `ole or `ore ; one occurs in the first quarter of the moon and the other in the third quarter. The same name is given to the prevailing wind during those phases. The words mean “nothing,” because those days were considered bad for fishing as well as planting.

Parts of Whorf ’s hypothesis, known as linguistic relativity , were controversial from the beginning, and still are among some linguists. Yet Whorf ’s ideas now form the basis for an entire sub-field of cultural anthropology: cognitive or psychological anthropology. A number of studies have been done that support Whorf ’s ideas. Linguist George Lakoff ’s work looks at the pervasive existence of metaphors in everyday speech that can be said to predispose a speaker’s world view and attitudes on a variety of human experiences. A metaphor is an expression in which one kind of thing is understood and experienced in terms of another entirely unrelated thing; the metaphors in a language can reveal aspects of the culture of its speakers. Take, for example, the concept of an argument. In logic and philosophy, an argument is a discussion involving differing points of view, or a debate. But the conceptual metaphor in American culture can be stated as ARGUMENT IS WAR. This metaphor is reflected in many expressions of the everyday language of American speakers: I won the argument. He shot down every point I made. They attacked every argument we made. Your point is right on target . I had a fight with my boyfriend last night. In other words, we use words appropriate for discussing war when we talk about arguments, which are certainly not real war. But we actually think of arguments as a verbal battle that often involve anger, and even violence, which then structures how we argue.

To illustrate that this concept of argument is not universal, Lakoff suggests imagining a culture where an argument is not something to be won or lost, with no strategies for attacking or defending, but rather as a dance where the dancers’ goal is to perform in an artful, pleasing way. No anger or violence would occur or even be relevant to speakers of this language, because the metaphor for that culture would be ARGUMENT IS DANCE.

3.1 Adapted from Perspectives , Language ( Linda Light, 2017 )

You can either watch the video, How Language Shapes the Way We Think, by linguist Lera Boroditsky, or read the script below.

Watch the video: How Language Shapes the Way We Think ( Boroditsky, 2018)

There are about 7,000 languages spoken around the world—and they all have different sounds, vocabularies, and structures. But do they shape the way we think? Cognitive scientist Lera Boroditsky shares examples of language—from an Aboriginal community in Australia that uses cardinal directions instead of left and right to the multiple words for blue in Russian—that suggest the answer is a resounding yes. “The beauty of linguistic diversity is that it reveals to us just how ingenious and how flexible the human mind is,” Boroditsky says. “Human minds have invented not one cognitive universe, but 7,000.”

Video transcript:

So, I’ll be speaking to you using language ... because I can. This is one these magical abilities that we humans have. We can transmit really complicated thoughts to one another. So what I’m doing right now is, I’m making sounds with my mouth as I’m exhaling. I’m making tones and hisses and puffs, and those are creating air vibrations in the air. Those air vibrations are traveling to you, they’re hitting your eardrums, and then your brain takes those vibrations from your eardrums and transforms them into thoughts. I hope.

I hope that’s happening. So because of this ability, we humans are able to transmit our ideas across vast reaches of space and time. We’re able to transmit knowledge across minds. I can put a bizarre new idea in your mind right now. I could say, “Imagine a jellyfish waltzing in a library while thinking about quantum mechanics.”

Now, if everything has gone relatively well in your life so far, you probably haven’t had that thought before.

But now I’ve just made you think it, through language.

Now of course, there isn’t just one language in the world, there are about 7,000 languages spoken around the world. And all the languages differ from one another in all kinds of ways. Some languages have different sounds, they have different vocabularies, and they also have different structures—very importantly, different structures. That begs the question: Does the language we speak shape the way we think? Now, this is an ancient question. People have been speculating about this question forever. Charlemagne, Holy Roman emperor, said, “To have a second language is to have a second soul”—strong statement that language crafts reality. But on the other hand, Shakespeare has Juliet say, “What’s in a name? A rose by any other name would smell as sweet.” Well, that suggests that maybe language doesn’t craft reality.

These arguments have gone back and forth for thousands of years. But until recently, there hasn’t been any data to help us decide either way. Recently, in my lab and other labs around the world, we’ve started doing research, and now we have actual scientific data to weigh in on this question.

So let me tell you about some of my favorite examples. I’ll start with an example from an Aboriginal community in Australia that I had a chance to work with. These are the Kuuk Thaayorre people. They live in Pormpuraaw at the very west edge of Cape York. What’s cool about Kuuk Thaayorre is, in Kuuk Thaayorre, they don’t use words like “left” and “right,” and instead, everything is in cardinal directions: north, south, east, and west. And when I say everything, I really mean everything. You would say something like, “Oh, there’s an ant on your southwest leg.” Or, “Move your cup to the north-northeast a little bit.” In fact, the way that you say “hello” in Kuuk Thaayorre is you say, “Which way are you going?” And the answer should be, “North-northeast in the far distance. How about you?”

So imagine as you’re walking around your day, every person you greet, you have to report your heading direction.

But that would actually get you oriented pretty fast, right? Because you literally couldn’t get past “hello,” if you didn’t know which way you were going. In fact, people who speak languages like this stay oriented really well. They stay oriented better than we used to think humans could. We used to think that humans were worse than other creatures because of some biological excuse: “Oh, we don’t have magnets in our beaks or in our scales.” No; if your language and your culture trains you to do it, actually, you can do it. There are humans around the world who stay oriented really well.

And just to get us in agreement about how different this is from the way we do it, I want you all to close your eyes for a second and point southeast.

Keep your eyes closed. Point. OK, so you can open your eyes. I see you guys pointing there, there, there, there, there ... I don’t know which way it is myself—

You have not been a lot of help.

So let’s just say the accuracy in this room was not very high. This is a big difference in cognitive ability across languages, right? Where one group—very distinguished group like you guys—doesn’t know which way is which, but in another group, I could ask a five-year-old and they would know.

There are also really big differences in how people think about time. So here I have pictures of my grandfather at different ages. And if I ask an English speaker to organize time, they might lay it out this way, from left to right. This has to do with writing direction. If you were a speaker of Hebrew or Arabic, you might do it going in the opposite direction, from right to left.

But how would the Kuuk Thaayorre, this Aboriginal group I just told you about, do it? They don’t use words like “left” and “right.” Let me give you hint. When we sat people facing south, they organized time from left to right. When we sat them facing north, they organized time from right to left. When we sat them facing east, time came towards the body. What’s the pattern? East to west, right? So for them, time doesn’t actually get locked on the body at all, it gets locked on the landscape. So for me, if I’m facing this way, then time goes this way, and if I’m facing this way, then time goes this way. I’m facing this way, time goes this way— very egocentric of me to have the direction of time chase me around every time I turn my body. For the Kuuk Thaayorre, time is locked on the landscape. It’s a dramatically different way of thinking about time.

Here’s another really smart human trait. Suppose I ask you how many penguins are there. Well, I bet I know how you’d solve that problem if you solved it. You went, “One, two, three, four, five, six, seven, eight.” You counted them. You named each one with a number, and the last number you said was the number of penguins. This is a little trick that you’re taught to use as kids. You learn the number list and you learn how to apply it. A little linguistic trick. Well, some languages don’t do this, because some languages don’t have exact number words. They’re languages that don’t have a word like “seven” or a word like “eight.” In fact, people who speak these languages don’t count, and they have trouble keeping track of exact quantities. So, for example, if I ask you to match this number of penguins to the same number of ducks, you would be able to do that by counting. But folks who don’t have that linguistic trait can’t do that.

Languages also differ in how they divide up the color spectrum—the visual world. Some languages have lots of words for colors, some have only a couple words, “light” and “dark.” And languages differ in where they put boundaries between colors. So, for example, in English, there’s a word for blue that covers all of the colors that you can see on the screen, but in Russian, there isn’t a single word. Instead, Russian speakers have to differentiate between light blue, goluboy , and dark blue, siniy . So Russians have this lifetime of experience of, in language, distinguishing these two colors. When we test people’s ability to perceptually discriminate these colors, what we find is that Russian speakers are faster across this linguistic boundary. They’re faster to be able to tell the difference between a light and a dark blue. And when you look at people’s brains as they’re looking at colors—say you have colors shifting slowly from light to dark blue—the brains of people who use different words for light and dark blue will give a surprised reaction as the colors shift from light to dark, as if, “Ooh, something has categorically changed,” whereas the brains of English speakers, for example, that don’t make this categorical distinction, don’t give that surprise, because nothing is categorically changing.

Languages have all kinds of structural quirks. This is one of my favorites. Lots of languages have grammatical gender; so every noun gets assigned a gender, often masculine or feminine. And these genders differ across languages. So, for example, the sun is feminine in German but masculine in Spanish, and the moon, the reverse. Could this actually have any consequence for how people think? Do German speakers think of the sun as somehow more female-like, and the moon somehow more male-like? Actually, it turns out that’s the case. So if you ask German and Spanish speakers to, say, describe a bridge, like the one here—“bridge” happens to be grammatically feminine in German, grammatically masculine in Spanish—German speakers are more likely to say bridges are “beautiful,” “elegant,” and stereotypically feminine words. Whereas Spanish speakers will be more likely to say they’re “strong” or “long,” these masculine words.

Languages also differ in how they describe events, right? You take an event like this, an accident. In English, it’s fine to say, “He broke the vase.” In a language like Spanish, you might be more likely to say, “The vase broke,” or “The vase broke itself.” If it’s an accident, you wouldn’t say that someone did it. In English, quite weirdly, we can even say things like, “I broke my arm.” Now, in lots of languages, you couldn’t use that construction unless you are a lunatic and you went out looking to break your arm—[laughter] and you succeeded. If it was an accident, you would use a different construction.

Now, this has consequences. So, people who speak different languages will pay attention to different things, depending on what their language usually requires them to do. So we show the same accident to English speakers and Spanish speakers, English speakers will remember who did it, because English requires you to say, “He did it; he broke the vase.” Whereas Spanish speakers might be less likely to remember who did it if it’s an accident, but they’re more likely to remember that it was an accident. They’re more likely to remember the intention. So, two people watch the same event, witness the same crime, but end up remembering different things about that event. This has implications, of course, for eyewitness testimony. It also has implications for blame and punishment. So if you take English speakers and I just show you someone breaking a vase, and I say, “He broke the vase,” as opposed to “The vase broke,” even though you can witness it yourself, you can watch the video, you can watch the crime against the vase, you will punish someone more, you will blame someone more if I just said, “He broke it,” as opposed to, “It broke.” The language guides our reasoning about events.

Now, I’ve given you a few examples of how language can profoundly shape the way we think, and it does so in a variety of ways. So language can have big effects, like we saw with space and time, where people can lay out space and time in completely different coordinate frames from each other. Language can also have really deep effects—that’s what we saw with the case of number. Having count words in your language, having number words, opens up the whole world of mathematics. Of course, if you don’t count, you can’t do algebra, you can’t do any of the things that would be required to build a room like this or make this broadcast, right? This little trick of number words gives you a stepping stone into a whole cognitive realm.

Language can also have really early effects, what we saw in the case of color. These are really simple, basic, perceptual decisions. We make thousands of them all the time, and yet, language is getting in there and fussing even with these tiny little perceptual decisions that we make. Language can have really broad effects. So the case of grammatical gender may be a little silly, but at the same time, grammatical gender applies to all nouns. That means language can shape how you’re thinking about anything that can be named by a noun. That’s a lot of stuff.

And finally, I gave you an example of how language can shape things that have personal weight to us—ideas like blame and punishment or eyewitness memory. These are important things in our daily lives.

Now, the beauty of linguistic diversity is that it reveals to us just how ingenious and how flexible the human mind is. Human minds have invented not one cognitive universe, but 7,000—there are 7,000 languages spoken around the world. And we can create many more—languages, of course, are living things, things that we can hone and change to suit our needs. The tragic thing is that we’re losing so much of this linguistic diversity all the time. We’re losing about one language a week, and by some estimates, half of the world’s languages will be gone in the next hundred years. And the even worse news is that right now, almost everything we know about the human mind and human brain is based on studies of usually American English-speaking undergraduates at universities. That excludes almost all humans. Right? So what we know about the human mind is actually incredibly narrow and biased, and our science has to do better.

I want to leave you with this final thought. I’ve told you about how speakers of different languages think differently, but of course, that’s not about how people elsewhere think. It’s about how you think. It’s how the language that you speak shapes the way that you think. And that gives you the opportunity to ask, “Why do I think the way that I do?” “How could I think differently?” And also, “What thoughts do I wish to create?”

Thank you very much.

Read the following text on what lexical differences between language can tell us about those languages’ cultures.

  • Subject List
  • Take a Tour
  • For Authors
  • Subscriber Services
  • Publications
  • African American Studies
  • African Studies
  • American Literature
  • Anthropology
  • Architecture Planning and Preservation
  • Art History
  • Atlantic History
  • Biblical Studies
  • British and Irish Literature
  • Childhood Studies
  • Chinese Studies
  • Cinema and Media Studies
  • Communication
  • Criminology
  • Environmental Science
  • Evolutionary Biology
  • International Law
  • International Relations
  • Islamic Studies
  • Jewish Studies
  • Latin American Studies
  • Latino Studies

Linguistics

  • Literary and Critical Theory
  • Medieval Studies
  • Military History
  • Political Science
  • Public Health
  • Renaissance and Reformation
  • Social Work
  • Urban Studies
  • Victorian Literature
  • Browse All Subjects

How to Subscribe

  • Free Trials

In This Article Expand or collapse the "in this article" section Innateness

Introduction.

  • Handbooks and Edited Collections
  • Bibliographies
  • The Public Broadcasting Service
  • Early Modern Philosophy
  • Contemporary Approaches
  • Evolutionary and Biological Influences
  • Nativist Rebuttals
  • Non-Nativist Approaches in Psycholinguistics

Related Articles Expand or collapse the "related articles" section about

About related articles close popup.

Lorem Ipsum Sit Dolor Amet

Vestibulum ante ipsum primis in faucibus orci luctus et ultrices posuere cubilia Curae; Aliquam ligula odio, euismod ut aliquam et, vestibulum nec risus. Nulla viverra, arcu et iaculis consequat, justo diam ornare tellus, semper ultrices tellus nunc eu tellus.

  • Critical Periods
  • Cross-Language Speech Perception and Production
  • Developmental Pragmatics
  • Embodiment and Language
  • Philosophy of Language
  • Sentence Processing in Monolingual and Bilingual Speakers

Other Subject Areas

Forthcoming articles expand or collapse the "forthcoming articles" section.

  • Cognitive Grammar
  • Edward Sapir
  • Find more forthcoming articles...
  • Export Citations
  • Share This Facebook LinkedIn Twitter

Innateness by Yarden Kedar LAST REVIEWED: 30 August 2022 LAST MODIFIED: 30 September 2013 DOI: 10.1093/obo/9780199772810-0071

The concept of innateness ( innate is first recorded in the period 1375–1425; from Latin innātus  “inborn”) relates to types of behavior and knowledge that are present in the organism since birth (in fact, since fertilization), prior to any sensory experience with the environment. The term has been applied to two general types of qualities. The first consists of instinctive and inflexible reflexes and behaviors, which are apparent in survival, mating, and rearing activities. The other relates to cognition, with certain concepts, ideas, propositions, and particular ways of mental computation suggested to be part of one’s biological makeup. While both types of innatism have a long history in human philosophy and science (e.g., Plato and Descartes), some bias appears to exist in favor of claims for inherent behavioral traits, which are typically accepted when satisfactory empirical evidence is provided. One famous example is Lorenz’s demonstration of imprinting, a natural phenomenon that obeys a predetermined mechanism and schedule (Lorenz’s incubator-hatched goslings imprinted on his boots, the first moving object they encountered). Likewise, there seems to be little controversy in regard to predetermined ways of organizing sensory information, as is the case with the detection and classification of shapes and colors by the mind. In contrast, the idea that certain types of abstract knowledge may be part of an organism’s biological endowment (i.e., not learned) is typically faced with a greater sense of skepticism, and touches on a fundamental question in epistemological philosophy: Can reason be based (to a certain extent) on a priori knowledge—that is, knowledge that precedes and is independent of experience? The most influential and controversial claim for such innate knowledge in modern science is Chomsky’s breakthrough nativist theory of Universal Grammar in language and the famous “argument from the poverty of the stimulus.” The main Chomskyan hypothesis is that all human beings share a preprogrammed linguistic infrastructure consisting of a finite collection of rules, which in principle may generate (through combination or transformation) an infinite number of (only) grammatical sentences. Thus the innate grammatical system constrains and structures the acquisition and use of all natural languages.

Several textbooks are relevant to the concept of innateness and its use in the theoretical and empirical psycholinguistic literature. However, in most cases these resources do not address innateness per se but rather as part of another topic. Some leading textbooks include Levitin 2010 on cognitive psychology, Sigelman and Rider 2008 on human development, and Bornstein and Lamb 2011 on developmental psychology. A comprehensive neuroscience perspective is given in Gazzaniga 2009 . The following textbooks focus on the acquisition of language, with specific reference to innate factors that influence this developmental process: Berko Gleason and Bernstein Ratner 2012 , with special attention given to language disorders and new findings on the brain and language; Clark 2009 , which presents a clear argument against an assumption of innate and language-specific mechanisms in the acquisition process; Lust 2006 , taking a generative approach, with many mentions of the innateness question in psycholinguistics; and Saxton 2010 , emphasizing the importance of the nature-nurture debate in the cognitive and language sciences.

Berko Gleason, Jean, and Nan Bernstein Ratner, eds. 2012. The development of language . 8th ed. Boston: Pearson Education.

Covers language change through the entire life span, including the aging process. Details the development of each of the main components of language (syntax, morphology, semantics, phonology, and pragmatics).

Bornstein, M. H., and M. E. Lamb, eds. 2011. Cognitive development: An advanced textbook . New York: Psychology Press.

An essential textbook by two prominent scholars in developmental science, which surveys contemporary theories, methods, and empirical findings in the field, including a specific section on language.

Clark, E. V. 2009. First language acquisition . Cambridge, UK: Cambridge Univ. Press.

DOI: 10.1017/CBO9780511806698

Presents a stage-based, bottom-up, learning view on language development in monolingual and bilingual children. Addresses language comprehension and use from preverbal stages to advanced competencies such as debating, instructing, and storytelling.

Gazzaniga, M. 2009. The cognitive neurosciences . 4th ed. Cambridge, MA: MIT Press.

An extensive resource edited by one of the most knowledgeable scholars in cognitive neuroscience. Discusses and gives examples of the effect of brain structures and functions on behavior in many respects, including language.

Levitin, D. J. 2010. Foundations of cognitive psychology: Core readings . 2d ed. Boston: Allyn & Bacon.

A comprehensive yet entertaining overview of cognitive psychology. Addresses some aspects of cognition that are not typically discussed in introductory works, most notably the relation between music and cognition.

Lust, B. 2006. Child language: Acquisition and growth . New York: Cambridge Univ. Press.

DOI: 10.1017/CBO9780511803413

A most valuable overview of past and present works in the study of child language. An excellent resource for both novices and experts in the field.

Saxton, M. 2010. Child language: Acquisition and development . London: SAGE.

Written mainly for psychology students and does not assume prior linguistic background. Calls for a middle-ground approach in the nature-nurture debate in regard to language.

Sigelman, C. K., and E. A. Rider. 2008. Life-span human development . Belmont, CA: Cengage Learning.

Covers the human life span from infancy to adulthood. Includes special sections on cognition, with a focus on the complexity of nature and nurture interactions in development.

back to top

Users without a subscription are not able to see the full content on this page. Please subscribe or login .

Oxford Bibliographies Online is available by subscription and perpetual access to institutions. For more information or to contact an Oxford Sales Representative click here .

  • About Linguistics »
  • Meet the Editorial Board »
  • Acceptability Judgments
  • Accessibility Theory in Linguistics
  • Acquisition, Second Language, and Bilingualism, Psycholin...
  • Adpositions
  • African Linguistics
  • Afroasiatic Languages
  • Algonquian Linguistics
  • Altaic Languages
  • Ambiguity, Lexical
  • Analogy in Language and Linguistics
  • Animal Communication
  • Applicatives
  • Applied Linguistics, Critical
  • Arawak Languages
  • Argument Structure
  • Artificial Languages
  • Australian Languages
  • Austronesian Linguistics
  • Auxiliaries
  • Balkans, The Languages of the
  • Baudouin de Courtenay, Jan
  • Berber Languages and Linguistics
  • Bilingualism and Multilingualism
  • Biology of Language
  • Borrowing, Structural
  • Caddoan Languages
  • Caucasian Languages
  • Celtic Languages
  • Celtic Mutations
  • Chomsky, Noam
  • Chumashan Languages
  • Classifiers
  • Clauses, Relative
  • Clinical Linguistics
  • Cognitive Linguistics
  • Colonial Place Names
  • Comparative Reconstruction in Linguistics
  • Comparative-Historical Linguistics
  • Complementation
  • Complexity, Linguistic
  • Compositionality
  • Compounding
  • Computational Linguistics
  • Conditionals
  • Conjunctions
  • Connectionism
  • Consonant Epenthesis
  • Constructions, Verb-Particle
  • Contrastive Analysis in Linguistics
  • Conversation Analysis
  • Conversation, Maxims of
  • Conversational Implicature
  • Cooperative Principle
  • Coordination
  • Creoles, Grammatical Categories in
  • Cyberpragmatics
  • Default Semantics
  • Definiteness
  • Dementia and Language
  • Dene (Athabaskan) Languages
  • Dené-Yeniseian Hypothesis, The
  • Dependencies
  • Dependencies, Long Distance
  • Derivational Morphology
  • Determiners
  • Dialectology
  • Distinctive Features
  • Dravidian Languages
  • Endangered Languages
  • English as a Lingua Franca
  • English, Early Modern
  • English, Old
  • Eskimo-Aleut
  • Euphemisms and Dysphemisms
  • Evidentials
  • Exemplar-Based Models in Linguistics
  • Existential
  • Existential Wh-Constructions
  • Experimental Linguistics
  • Fieldwork, Sociolinguistic
  • Finite State Languages
  • First Language Attrition
  • Formulaic Language
  • Francoprovençal
  • French Grammars
  • Gabelentz, Georg von der
  • Genealogical Classification
  • Generative Syntax
  • Genetics and Language
  • Grammar, Categorial
  • Grammar, Construction
  • Grammar, Descriptive
  • Grammar, Functional Discourse
  • Grammars, Phrase Structure
  • Grammaticalization
  • Harris, Zellig
  • Heritage Languages
  • History of Linguistics
  • History of the English Language
  • Hmong-Mien Languages
  • Hokan Languages
  • Humor in Language
  • Hungarian Vowel Harmony
  • Idiom and Phraseology
  • Imperatives
  • Indefiniteness
  • Indo-European Etymology
  • Inflected Infinitives
  • Information Structure
  • Interface Between Phonology and Phonetics
  • Interjections
  • Iroquoian Languages
  • Isolates, Language
  • Jakobson, Roman
  • Japanese Word Accent
  • Jones, Daniel
  • Juncture and Boundary
  • Khoisan Languages
  • Kiowa-Tanoan Languages
  • Kra-Dai Languages
  • Labov, William
  • Language Acquisition
  • Language and Law
  • Language Contact
  • Language Documentation
  • Language, Embodiment and
  • Language for Specific Purposes/Specialized Communication
  • Language, Gender, and Sexuality
  • Language Geography
  • Language Ideologies and Language Attitudes
  • Language in Autism Spectrum Disorders
  • Language Nests
  • Language Revitalization
  • Language Shift
  • Language Standardization
  • Language, Synesthesia and
  • Languages of Africa
  • Languages of the Americas, Indigenous
  • Languages of the World
  • Learnability
  • Lexical Access, Cognitive Mechanisms for
  • Lexical Semantics
  • Lexical-Functional Grammar
  • Lexicography
  • Lexicography, Bilingual
  • Linguistic Accommodation
  • Linguistic Anthropology
  • Linguistic Areas
  • Linguistic Landscapes
  • Linguistic Prescriptivism
  • Linguistic Profiling and Language-Based Discrimination
  • Linguistic Relativity
  • Linguistics, Educational
  • Listening, Second Language
  • Literature and Linguistics
  • Machine Translation
  • Maintenance, Language
  • Mande Languages
  • Mass-Count Distinction
  • Mathematical Linguistics
  • Mayan Languages
  • Mental Health Disorders, Language in
  • Mental Lexicon, The
  • Mesoamerican Languages
  • Minority Languages
  • Mixed Languages
  • Mixe-Zoquean Languages
  • Modification
  • Mon-Khmer Languages
  • Morphological Change
  • Morphology, Blending in
  • Morphology, Subtractive
  • Munda Languages
  • Muskogean Languages
  • Nasals and Nasalization
  • Niger-Congo Languages
  • Non-Pama-Nyungan Languages
  • Northeast Caucasian Languages
  • Oceanic Languages
  • Papuan Languages
  • Penutian Languages
  • Phonetics, Acoustic
  • Phonetics, Articulatory
  • Phonological Research, Psycholinguistic Methodology in
  • Phonology, Computational
  • Phonology, Early Child
  • Policy and Planning, Language
  • Politeness in Language
  • Positive Discourse Analysis
  • Possessives, Acquisition of
  • Pragmatics, Acquisition of
  • Pragmatics, Cognitive
  • Pragmatics, Computational
  • Pragmatics, Cross-Cultural
  • Pragmatics, Developmental
  • Pragmatics, Experimental
  • Pragmatics, Game Theory in
  • Pragmatics, Historical
  • Pragmatics, Institutional
  • Pragmatics, Second Language
  • Pragmatics, Teaching
  • Prague Linguistic Circle, The
  • Presupposition
  • Psycholinguistics
  • Quechuan and Aymaran Languages
  • Reading, Second-Language
  • Reciprocals
  • Reduplication
  • Reflexives and Reflexivity
  • Register and Register Variation
  • Relevance Theory
  • Representation and Processing of Multi-Word Expressions in...
  • Salish Languages
  • Sapir, Edward
  • Saussure, Ferdinand de
  • Second Language Acquisition, Anaphora Resolution in
  • Semantic Maps
  • Semantic Roles
  • Semantic-Pragmatic Change
  • Semantics, Cognitive
  • Sign Language Linguistics
  • Sociolinguistics
  • Sociolinguistics, Variationist
  • Sociopragmatics
  • Sound Change
  • South American Indian Languages
  • Specific Language Impairment
  • Speech, Deceptive
  • Speech Perception
  • Speech Production
  • Speech Synthesis
  • Switch-Reference
  • Syntactic Change
  • Syntactic Knowledge, Children’s Acquisition of
  • Tense, Aspect, and Mood
  • Text Mining
  • Tone Sandhi
  • Transcription
  • Transitivity and Voice
  • Translanguaging
  • Translation
  • Trubetzkoy, Nikolai
  • Tucanoan Languages
  • Tupian Languages
  • Usage-Based Linguistics
  • Uto-Aztecan Languages
  • Valency Theory
  • Verbs, Serial
  • Vocabulary, Second Language
  • Voice and Voice Quality
  • Vowel Harmony
  • Whitney, William Dwight
  • Word Classes
  • Word Formation in Japanese
  • Word Recognition, Spoken
  • Word Recognition, Visual
  • Word Stress
  • Writing, Second Language
  • Writing Systems
  • Zapotecan Languages
  • Privacy Policy
  • Cookie Policy
  • Legal Notice
  • Accessibility

Powered by:

  • [66.249.64.20|185.147.128.134]
  • 185.147.128.134

Cambridge Dictionary

  • Cambridge Dictionary +Plus

Meaning of hypothesis in English

Your browser doesn't support HTML5 audio

  • abstraction
  • afterthought
  • anthropocentrism
  • anti-Darwinian
  • exceptionalism
  • foundation stone
  • great minds think alike idiom
  • non-dogmatic
  • non-empirical
  • non-material
  • non-practical
  • social Darwinism
  • supersensible
  • the domino theory

hypothesis | American Dictionary

Hypothesis | business english, examples of hypothesis, translations of hypothesis.

Get a quick, free translation!

{{randomImageQuizHook.quizId}}

Word of the Day

to try to persuade a customer who is already buying something to buy more, or to buy something more expensive

Searching out and tracking down: talking about finding or discovering things

Searching out and tracking down: talking about finding or discovering things

hypothesis definition in linguistics

Learn more with +Plus

  • Recent and Recommended {{#preferredDictionaries}} {{name}} {{/preferredDictionaries}}
  • Definitions Clear explanations of natural written and spoken English English Learner’s Dictionary Essential British English Essential American English
  • Grammar and thesaurus Usage explanations of natural written and spoken English Grammar Thesaurus
  • Pronunciation British and American pronunciations with audio English Pronunciation
  • English–Chinese (Simplified) Chinese (Simplified)–English
  • English–Chinese (Traditional) Chinese (Traditional)–English
  • English–Dutch Dutch–English
  • English–French French–English
  • English–German German–English
  • English–Indonesian Indonesian–English
  • English–Italian Italian–English
  • English–Japanese Japanese–English
  • English–Norwegian Norwegian–English
  • English–Polish Polish–English
  • English–Portuguese Portuguese–English
  • English–Spanish Spanish–English
  • English–Swedish Swedish–English
  • Dictionary +Plus Word Lists
  • English    Noun
  • American    Noun
  • Business    Noun
  • Translations
  • All translations

To add hypothesis to a word list please sign up or log in.

Add hypothesis to one of your lists below, or create a new one.

{{message}}

Something went wrong.

There was a problem sending your report.

The Affective Filter Hypothesis: Definition and Criticism

January 23, 2018, 9:00 am

The Affective Filter Hypothesis: Definition and Criticism

Linguist and educator Stephen Krashen proposed the Monitor Model, his theory of second language acquisition, in Principles and practice in second language acquisition as published in 1982. According to the Monitor Model, five hypotheses account for the acquisition of a second language:

  • Acquisition-learning hypothesis
  • Natural order hypothesis
  • Monitor hypothesis
  • Input hypothesis
  • Affective filter hypothesis

However, in spite of the popularity and influence of the Monitor Model, the five hypotheses are not without criticism. The following sections offer a description of the fifth and final hypothesis of the theory, the affective filter hypothesis, as well as the major criticism by other linguistics and educators surrounding the hypothesis.

Definition of the Affective Filter Hypothesis

The fifth hypothesis, the affective filter hypothesis, accounts for the influence of affective factors on second language acquisition. Affect refers to non-linguistic variables such as motivation, self-confidence, and anxiety. According to the affective filter hypothesis, affect effects acquisition, but not learning, by facilitating or preventing comprehensible input from reaching the language acquisition device. In other words, affective variables such as fear, nervousness, boredom, and resistance to change can effect the acquisition of a second language by preventing information about the second language from reaching the language areas of the mind.

Furthermore, when the affective filter blocks comprehensible input, acquisition fails or occurs to a lesser extent then when the affective filter supports the intake of comprehensible input. The affective filter, therefore, accounts for individual variation in second language acquisition. Second language instruction can and should work to minimize the effects of the affective filter.

Criticism of the Affective Filter Hypothesis

The final critique of Krashen’s Monitor Model questions the claim of the affective filter hypothesis that affective factors alone account for individual variation in second language acquisition. First, Krashen claims that children lack the affective filter that causes most adult second language learners to never completely master their second language. Such a claim fails to withstand scrutiny because children also experience differences in non-linguistic variables such as motivation, self-confidence, and anxiety that supposedly account for child-adult differences in second language learning.

Furthermore, evidence in the form of adult second language learners who acquire a second language to a native-like competence except for a single grammatical feature problematizes the claim that an affective filter prevents comprehensible input from reaching the language acquisition device. As Manmay Zafar asks, “How does the filter determine which parts of language are to be screened in/out?” In other words, the affective filter hypothesis fails to answer the most important question about affect alone accounting for individual variation in second language acquisition.

Although the Monitor Model has been influential in the field of second language acquisition, the fifth and final hypothesis, the affective filter hypothesis, has not been without criticism as evidenced by the critiques offered by other linguists and educators in the field.

Gass, Susan M. & Larry Selinker. 2008. Second language acquisition: An introductory course , 3rd edn. New York: Routledge. Gregg, Kevin R. 1984. Krashen’s monitor and Occam’s razor. Applied Linguistics 5(2). 79-100. Krashen, Stephen D. 1982. Principles and practice in second language acquisition . Oxford: Pergamon. http://www.sdkrashen.com/Principles_and_Practice/Principles_and_Practice.pdf. Lightbrown, Patsy M. & Nina Spada. 2006. How languages are learned , 3rd edn. Oxford: Oxford University Press. Zafar, Manmay. 2009. Monitoring the ‘monitor’: A critique of Krashen’s five hypotheses. Dhaka University Journal of Linguistics 2(4). 139-146.

affective filter hypothesis language acquisition language learning monitor model

The Input Hypothesis: Definition and Criticism

The Input Hypothesis: Definition and Criticism

The Postpositional Complement in English Grammar

The Postpositional Complement in English Grammar

COMMENTS

  1. The Sapir-Whorf Hypothesis: How Language Influences How We Express

    Linguistic Relativity in Psychology. The Sapir-Whorf Hypothesis, also known as linguistic relativity, refers to the idea that the language a person speaks can influence their worldview, thought, and even how they experience and understand the world. While more extreme versions of the hypothesis have largely been discredited, a growing body of ...

  2. Sapir-Whorf hypothesis (Linguistic Relativity Hypothesis)

    The Sapir-Whorf hypothesis states that people experience the world based on the structure of their language, and that linguistic categories shape and limit cognitive processes. It proposes that differences in language affect thought, perception, and behavior, so speakers of different languages think and act differently.

  3. Philosophy of Linguistics > Whorfianism (Stanford Encyclopedia of

    Linguistic anthropologists have explicitly taken up the task of defending a famous claim associated with Sapir that connects linguistic variation to differences in thinking and cognition more generally. The claim is very often referred to as the Sapir-Whorf Hypothesis (though this is a largely infelicitous label, as we shall see).

  4. Definition and History of the Sapir-Whorf Hypothesis

    The Sapir-Whorf Hypothesis Linguistic Theory. Benjamin Whorf argued that "we dissect nature along lines laid down by our native languages". The Sapir-Whorf hypothesis is the linguistic theory that the semantic structure of a language shapes or limits the ways in which a speaker forms conceptions of the world. It came about in 1929.

  5. Sapir-Whorf Hypothesis

    Sapir-Whorf Hypothesis. The Sapir-Whorf hypothesis, also known as the linguistic relativity hypothesis, refers to the proposal that the particular language one speaks influences the way one thinks about reality. Linguistic relativity stands in close relation to semiotic-level concerns with the general relation of language and thought, and ...

  6. Linguistic relativity

    Linguistic relativity. The idea of linguistic relativity, known also as the Whorf hypothesis, the Sapir-Whorf hypothesis ( / səˌpɪər ˈhwɔːrf / sə-PEER WHORF ), or Whorfianism, is a principle suggesting that the structure of a language influences its speakers' worldview or cognition, and thus individuals' languages determine or ...

  7. Whorfian Hypothesis

    The term Whorfian Hypothesis takes its name from Benjamin Lee Whorf (1876-1941) who claimed that the language one speaks influences one's thinking [ 7 ]. Whorf was an amateur linguist who studied with the anthropologist Edward Sapir in the 1920s and 1930s. The term Sapir-Whorf Hypothesis is also used to refer to their view that language ...

  8. Relativism > The Linguistic Relativity Hypothesis (Stanford

    The linguistic relativity hypothesis grained its widest audience through the work of Benjamin Lee Whorf, whose collected writings became something of a relativistic manifesto. Whorf presents a moving target, with most of his claims coming in both extreme and in more cautious forms. Debate continues about his considered views, but there is ...

  9. Philosophy of Linguistics

    Philosophy of Linguistics. Philosophy of linguistics is the philosophy of science as applied to linguistics. This differentiates it sharply from the philosophy of language, traditionally concerned with matters of meaning and reference. As with the philosophy of other special sciences, there are general topics relating to matters like ...

  10. Cognitive Linguistics and Linguistic Relativity

    Linguistic relativity (also known as the Sapir-Whorf Hypothesis) is a general cover term for the conjunction of two basic notions. The first noti ... In addition to linguistic relativity, his general interests include semantic typology, field/descriptive linguistics (South India), and the representation of events. Eric Pederson can be reached ...

  11. Linguistic Relativity

    Linguistic relativity, sometimes called the Whorfian hypothesis, posits that properties of language affect the structure and content of thought and thus the way humans perceive reality. A distinction is often made between strong Whorfian views, according to which the categories of thought are determined by language, and weak views, which argue ...

  12. 3.1: Linguistic Relativity- The Sapir-Whorf Hypothesis

    After completing this module, students will be able to: 1. Define the concept of linguistic relativity. 2. Differentiate linguistic relativity and linguistic determinism. 3. Define the Sapir-Whorf Hypothesis (against more pop-culture takes on it) and situate it in a broader theoretical context/history. 4.

  13. (PDF) The Sapir-Whorf Hypothesis

    The S apir-Whorf hypothesis, commonly referred to as the linguistic relativity hypothesis, explores the idea that the. language one uses affects how one perceives reality. J.A. Lucy, (2001) [1 ...

  14. The 'innateness hypothesis' and explanatory models in linguistics

    To begin, then, the IH is the hypothesis that the human brain is 'programmed' at birth in some quite specific and structured aspects of human natural language. The details of this programming are spelled out in some detail in 'Explanatory models in linguistics'. We should assume that the speaker has 'built in' a function which ...

  15. The Input Hypothesis: Definition and Criticism

    The Input Hypothesis: Definition and Criticism. Stephen Krashen is a linguist and educator who proposed the Monitor Model, a theory of second language acquisition, in Principles and practice in second language acquisition as published in 1982. According to the Monitor Model, five hypotheses account for the acquisition of a second language:

  16. Innateness

    The main Chomskyan hypothesis is that all human beings share a preprogrammed linguistic infrastructure consisting of a finite collection of rules, which in principle may generate (through combination or transformation) an infinite number of (only) grammatical sentences. Thus the innate grammatical system constrains and structures the ...

  17. 'Interlanguage' in: The Encyclopedia of Applied Linguistics

    Interlanguage (IL) refers to the linguistic system of learner language produced by adults when they attempt meaningful communication using a language they are in the process of learning. The construct of interlanguage was proposed by Larry Selinker in 1972 and stimulated the first research studies in the new field of second language acquisition ...

  18. Innateness hypothesis

    In linguistics, the innateness hypothesis, also known as the nativist hypothesis, holds that humans are born with at least some knowledge of linguistic structure.On this hypothesis, language acquisition involves filling in the details of an innate blueprint rather than being an entirely inductive process. The hypothesis is one of the cornerstones of generative grammar and related approaches in ...

  19. The Monitor Hypothesis: Definition and Criticism

    Definition of the Monitor Hypothesis. The third hypothesis, the monitor hypothesis, complements the acquisition-learning hypothesis by claiming that the only function of learning within second language acquisition is as an editor, or Monitor, for language use produced by the acquired system as well as to produce grammatical forms not yet acquired.

  20. The Acquisition-Learning Hypothesis: Definition and Criticism

    Acquisition occurs passively and unconsciously through implicit, informal, or natural learning, resulting in implicit knowledge and acquired competence of a language; in other words, to acquire a language is to "pick up" a language by relying on "feelings" of correctness rather than conscious knowledge of language rules.

  21. HYPOTHESIS

    HYPOTHESIS definition: 1. an idea or explanation for something that is based on known facts but has not yet been proved…. Learn more.

  22. The Natural Order Hypothesis: Definition and Criticism

    Definition of the Natural Order Hypothesis. The second hypothesis, the natural order hypothesis, argues that the acquisition of grammatical structures occurs in a predictable sequence. The natural order hypothesis applies to both first language acquisition and second language acquisition, but, although similar, the order of acquisition often ...

  23. Linguistic Relativity Definition & Examples

    The linguistic relativity hypothesis gives people a strong reason to try to preserve languages. Of the 7,000 languages spoken today, up to 6,300 are expected to die out by 2100.

  24. The Affective Filter Hypothesis: Definition and Criticism

    Definition of the Affective Filter Hypothesis. The fifth hypothesis, the affective filter hypothesis, accounts for the influence of affective factors on second language acquisition. Affect refers to non-linguistic variables such as motivation, self-confidence, and anxiety. According to the affective filter hypothesis, affect effects acquisition ...