Psycholinguistics: Crash Course Linguistics #11
Hi, I'm Taylor and welcome to Crash Course Linguistics!
Language happens thanks to the brain.
This spongy thing sitting in our skull is responsible for our abilities to acquire complex linguistic skills like phonology, semantics, syntax and reading.
It's what helps us write poems and compose texts, or get the meaning of a raised eyebrow or a string of random emojis.
Understanding the brain helps us understand how language is produced and interpreted.
We can point directly to the parts of our mouths and hands that are used to make language, but you can't reach in and feel your own brain.
Luckily, over the years, people have devised a whole range of ways of looking at the brain to figure out where and how language happens.
This is the field of psycholinguistics.
[THEME MUSIC]
Interest in the relationship between language and the brain really picked up in the 19th century.
Researchers observed people with language disorders and then looked at their brains after they died.
If damage to one part of the brain from a stroke, dementia, or a head injury correlated with a difference in the subject's ability to understand or produce language,
the scientists could predict there was a relationship there.
These kinds of injuries that affect our language abilities are known as aphasia.
Two famous kinds of aphasia discovered at the time are called Broca's aphasia and Wernicke's aphasia.
The areas of the brain related to each kind of aphasia therefore became known as Broca's area and Wernicke's area.
You may have heard about them in a Psychology or Anatomy class.
Broca's area is located around your left temple, and it was named after Paul Broca,
a 19th century French physician who noticed that people who'd been injured in this part of their brain acted in similar ways.
While they could still understand language, they could only produce maybe a handful of words, one at a time.
Broca's area affected their ability to speak or sign in a fluid, grammatical way.
In other words, it affected their ability to use morphosyntax.
Right around the same time Broca was making his discovery, German physician Carl Wernicke discovered that if a different part of the brain was injured, there was a totally different effect.
The people injured in this spot, located just above your left ear, tended to talk in a way that sounded fluent and grammatical but was nonsensical.
Wernicke's area is associated with the meaning of language.
But those 19th century studies were limited, and the brain is amazingly complex and flexible.
More recent research has found that some people can majorly damage Broca's area and never develop aphasia.
Other people can re-learn how to speak through extensive practice building on their ability to sing, which is controlled by a different part of the brain.
These newer studies help us understand neuroplasticity,
the ability of the brain to flexibly build and connect parts of the brain in response to injury or as part of learning.
And though the language areas are usually located on the left hemisphere of the brain,
some people's language areas are found predominantly in the right hemisphere, or spread across both sides
especially for left-handed or ambidextrous people.
So the relationship between language and the brain is even more complicated than we first thought.
Even now, errors and differences in language use can teach us about the different skills involved in language and how they're organized inside our minds.
We all sometimes forget a word that we know perfectly well, or accidentally swap words, parts of words, or idioms
what you might encounter as spoonerisms, tip of the tongue experiences or mixed metaphors.
These production errors tell us valuable things about how the mind handles language.
Like, you know when you just can't quite remember a word?
You know it, you almost have it, it's right there... you just can't retrieve it.
This phenomenon is known as a Tip of the Tongue experience,
and psycholinguists have found that people with a word on the tips of their tongues can often recall other information about it.
They might remember its meaning, its first letter, and sometimes how many syllables it has, but they can't quite recall the complete word.
Signed languages also have this phenomenon, which is known as Tip of the Fingers, naturally.
And signers experiencing Tips of the Fingers can also recall certain information about the sign they're seeking, especially the initial handshape and location of the hand.
They just can't recall the movement they'd need to complete the sign.
Tip of the Tongue and Finger experiences can show us how our thoughts are organized,
because we can have access to the first letter or initial hand position without having access to the remaining sounds or movement.
Knowing a word isn't a binary state of "yes" or "no" like a computer.
Our brains can also retain partial information.
Production errors are so useful that psycholinguists have techniques for trying to get people to make even more of them, so they can study those errors in a laboratory setting.
Psycholinguists can induce Tip of the Tongue or Finger experiences by asking people to translate words or to recall proper nouns.
Let's head to the Thought Bubble to try another psycholinguistic experiment right here!
In a moment two shapes are going to appear on the screen.
Let's decide which one is called kiki, and which one is called bouba.
Are you ready?
It's more than likely that you called the shape on the left ‘bouba' and the shape on the right ‘kiki'.
About nine out of every ten people make that choice.
Experiments have repeatedly shown that we think of voiceless stops like /k/ and high vowels like /i/ as small, sharp, crunchy, bright and spikey,
compared to voiced sounds like /b/ and rounded vowels like /u/, which are big, lumpy, dark and heavy, across many different languages.
The kiki/bouba experiment shows us that language is related to other sensory experiences.
Yes, words are still made up of arbitrary smaller parts, but it's not completely 100% arbitrary, and this mapping between senses can show up in some general tendencies for naming things across language.
One large study showed that words for "nose" were more likely to have a nasal sound like /m/ or /n/, across many different languages.
Marketers are very aware of these links
bouba wouldn't make a good name for a brand of crackers, but kiki would, because we generally want our crackers to be crispy.
But I'm sure bouba brand ice cream would taste much better — round and smooth and creamy.
Despite these general tendencies, there are also language-specific differences.
If your language doesn't have a /b/ or /u/ sound, you might not think of ‘bouba' as a possible word, so you might not associate it consistently with the blobby shape.
Different languages can also label the shapes differently depending on how their sound systems work.
Tone can influence how Mandarin speakers label these shapes.
The human brain doesn't completely separate our linguistic knowledge from other knowledge of the world, and experiments like the kiki/bouba test help show that.
Thanks, Thought Bubble!
Or should I say...thought bouba?
That's one kind of psycholinguistic experiment, but it's far from the only one.
Psycholinguists might use a priming experiment to test how closely words are related in the brain.
They “prime” the participant with one word and measure the speed of their responses to other words.
Say a subject is primed with the word “dog” and then has a faster response to “cat” than to other words, we might conclude “cat” and “dog” are more closely related in the brain.
We can also use gating experiments, where we measure how much of a word a participant needs to hear or see until they know I'm saying, say, “park”, instead of “part”.
Gating experiments show that sounds aren't always produced in discrete sequences like our alphabet makes them look.
Like, most English speakers will produce the /k/ sound in "cube" a little bit differently than /k/ sound in "calm".
Psycholinguists have even looked into such mysteries as whether swearing helps us manage pain.
In that study, psycholinguists compared how long people could hold their hand in a bucket of ice water when they were allowed to swear and when they were not.
When people were allowed to swear, they could hold their hand in the iced water for longer.
Huh! I'm definitely going to find a practical application for this!
Other ways of figuring out what's going on in the brain when we use language involve using various kinds of equipment.
Eyetracking studies try to figure out what we're thinking about based on what we're looking at.
Let's say we're reading a sentence like this one:
Now, "The rabbit crouched on the cushion" is a totally reasonable English sentence,
so that's where most people assume it's going at first.
But then when we get to the word "seemed", we need to re-evaluate.
That's where eyetracking shows that a lot of people look back at the earlier portion of the sentence to figure out what's going on--in this case, a structure more like
"The rabbit that was crouched on the cushion seemed friendly."
Misleading sentences like these are called garden path sentences,
because they seem to "lead you up a garden path" of one interpretation before you realize that there's actually a different structure going on.
Eyetracking and garden path sentences show us that we process sentences as we're experiencing them
we don't wait until we've seen or heard a whole sentence before starting to figure out what's going on.
Electro-encephalography or EEG records the electrical activity of neurons firing through a cap of small sensors on the scalp.
A psycholinguist might hook a person up to an EEG and say a sentence like, “my favourite ice cream is chocolate and socks.”
“Socks” is semantically unexpected information in a sentence that we assumed would be about food, so the brain reacts accordingly.
And an EEG is especially good at indicating when a surge in electricity happens.
So here it might map a kind of surge, known as N400, around 400 milliseconds after hearing “socks.”
EEGs are quiet and relatively affordable, but they can be disturbed even by small things like blinking.
Plus, they're not that great at mapping where things happen in the brain.
Functional magnetic resonance imaging, or fMRI, on the other hand, is relatively precise in locating brain activity
getting within a few millimeters of where the activity is happening.
It does this by measuring when there is increased oxygen in parts of the brain.
The more neural activity, or thinking, the more blood goes to the area of the brain, and that blood brings lots of oxygen to help those busy neurons.
For example, a psycholinguist might have someone learn, and recite back, a few words in a made-up language to see what happens in the brain when we try to learn a new language.
While fMRI is relatively precise in locating brain activity, it's less precise at when that activity is happening.
It only gets within a few seconds, while thoughts can happen in fractions of a second.
They're also very expensive and pretty dang uncomfortable to hang out in.
So there's sort of a tradeoff:
EEG machines are precise about time, but imprecise about space, whereas MRI machines are precise about space, but imprecise about time.
These machines, with their precise data and complex graphs, might seem like just the thing 19th century researchers like Broca and Wernicke needed to understand the link between the brain and language.
But really, we need to approach them with just as much caution as those older experiments.
There's still a lot of individual variation in how our brains get organized as we learn things,
and lots of psycholinguistics work has been done with small numbers of people who speak dominant languages like English.
So we only know a little about if and how being fluent in other languages affects what happens in the brain.
There's always more to learn.
See you next time, when we talk about how we learn language in the first place!
Thanks for watching this episode of Crash Course Linguistics.
If you want to help keep all Crash Course free for everybody, forever, you can join our community on Patreon.