YouTube Transcript: Sensation and Perception: Crash Course Psychology #5
Skip watching entire videos - get the full transcript, search for keywords, and copy with one click.
Share:
Video Transcript
Video Summary
Summary
Core Theme
This content explains the fundamental difference between sensation (receiving stimuli) and perception (interpreting stimuli), using the example of neurologist Oliver Sacks's prosopagnosia (face blindness) to illustrate how specific brain functions can be impaired while others remain intact. It further delves into the biological and psychological mechanisms of human vision and sensory perception.
Let me tell you about Oliver Sacks, the famous physician, professor and author of unusual
neurological case studies.
We’ll be looking at some of his fascinating research in future lessons, but for now, I
just want to talk about Sacks himself.
Although he possesses a brilliant and inquisitive mind, Dr. Sacks cannot do a simple thing that
your average toddler can.
He can’t recognize his own face in the mirror.
Sacks has a form of prosopagnosia, a neurological disorder that impairs a person’s ability
to perceive or recognize faces, also known as face blindness.
Last week we talked about how brain function is localized, and this is another peculiarly
excellent example of that.
Sacks can recognize his coffee cup on the shelf, but he can’t pick out his oldest
friend from a crowd, because the specific sliver of his brain responsible for facial
recognition is malfunctioning.
There’s nothing wrong with his vision.
The sense is intact.
The problem is with his perception, at least when it comes to recognizing faces.
Prosopagnosia is a good example of how sensing and perceiving are connected, but different.
Sensation is the bottom-up process by which our senses, like vision, hearing and smell,
receive and relay outside stimuli.
Perception, on the other hand, is the top-down way our brains organize and interpret that
information and put it into context.
So right now at this very moment, you’re probably receiving light from your screen
through your eyes, which will send the data of that sensation to your brain.
Perception meanwhile is your brain telling you that what you’re seeing is a diagram
explaining the difference between sensation and perception, which is pretty meta.
Now your brain is interpreting that light as a talking person, whom your brain might
additionally recognize as Hank.
[Intro]
We are constantly bombarded by stimuli even though we’re only aware of what our own
senses can pick up.
Like I can see and hear and feel and even smell this Corgi, but I can’t hunt using
sonar like a bat or hear a mole tunneling underground like an owl or see ultraviolet
and infrared light like a mantis shrimp.
I probably can’t even smell half of what you can smell.
No!
We have different senses.
Mwah mwah mwah mwah mwah.
Yeah.
There’s a lot to sense in the world, and not everybody needs to sense all the same
stuff.
So every animal has its limitations which we can talk about more precisely if we define
the Absolute Threshold of Sensation, the minimum stimulation needed to register a particular
stimulus, 50% of the time.
So if I play a tiny little beep in your ear and you tell me that you hear it fifty percent
of the times that I play it, that’s your absolute threshold of sensation.
We have to use a percentage because sometimes I'll play the beep and you’ll hear it and
sometimes you won’t even though it’s the exact same volume.
Why?
Because brains are complicated.
Detecting a weak sensory signal like that beep in daily life isn’t only about the
strength of the stimulus.
It’s also about your psychological state; your alertness and expectations in the moment.
This has to do with Signal Detection Theory, a model for predicting how and when a person
will detect a weak stimuli, partly based on context.
Exhausted new parents might hear their baby’s tiniest whimper, but not even register the
bellow of a passing train.
Their paranoid parent brains are so trained on their baby, it gives their senses a sort
of boosted ability, but only in relation to the subject of their attention.
Conversely, if you’re experiencing constant stimulation, your senses will adjust in a
process called sensory adaptation.
It is the reason that I have to check and see if my wallet is there if it’s in my
right pocket, but if I move it to my left pocket, it feels like a big uncomfortable
lump.
It’s also useful to be able to talk about our ability to detect the difference between
two stimuli.
I might go out at night and look up at the sky and, well, I know with my objective science
brain that no two stars have the exact same brightness, and yeah, I can tell with my eyeballs
that some stars are brighter than others, but other stars just look exactly the same
to me.
I can’t tell the difference in their brightness.
Are you done?
Is it time for your to go?
Gimme, gimme a kiiiissss.
Yes, yes.
Okay.
Good girl.
The point at which one can tell the difference is the difference threshold, but it’s not
linear.
Like.
if a tiny star is just a tiny bit brighter than another tiny star, I can tell.
But if a big star is that same tiny amount brighter than another big star, I won’t
be able to tell the difference.
This is important enough that we gave the guy who discovered it a law.
Weber’s Law says that we perceive differences on a logarithmic, not a linear scale.
It’s not the amount of change.
It’s the percentage change that matters.
Alright.
How about now we take a more in depth look at how one of our most powerful senses works?
Vision.
Your ability to see your face in the mirror is the result of a long but lightning quick
sequence of events.
Light bounces off your face and then off the mirror and then into your eyes, which take
in all that varied energy and transforms it into neural messages that your brain processes
and organizes into what you actually see, which is your face.
Or if you’re looking elsewhere, you could see a coffee cup or a Corgi or a scary clown
holding a tiny cream pie.
So how do we transform light waves into meaningful information?
Well, let’s start with the light itself.
What we humans see as light is only a small fraction of the full spectrum of electromagnetic
radiation that ranges from gamma to radio waves.
Now light has all kinds of fascinating characteristics that determine how we sense it, but for the
purposes of this topic, we’ll understand light as traveling in waves.
The wave’s wavelength and frequency determines their hue, and their amplitude determines
their intensity or brightness.
For instance a short wave has a high frequency.
Our eyes register short wavelengths with high frequencies as blueish colors while we see
long, low frequency wavelengths as reddish hues.
The way we register the brightness of a color, the contrast between the orange of a sherbet
and the orange of a construction cone has to do with the intensity or amount of energy
in a given light wave.
Which as we’ve just said is determined by its amplitude.
Greater amplitude means higher intensity, means brighter color.
Someone’s just told me that sherbet doesn’t- isn’t a word that exists.
His name is Michael Aranda and he’s a dumbhead.
Did you type it into the dictionary?
Type it into Google.
Ask Google about sherbet.
So sherbet is a thing.
So after taking this light in through the cornea and the pupil, it hits the transparent
disc behind the pupil: the lens, which focuses the light rays into specific images, and just
as you’d expect the lens to do, it projects these images onto the retina, the inner surface
of the eyeball that contains all the receptor cells that begin sensing that visual information.
Now your retinas don’t receive a full image like a movie being projected onto a screen.
It’s more like a bunch of pixel points of light energy that millions of receptors translate
into neural impulses and zip back into the brain.
These retinal receptors are called rods and cones.
Our rods detect gray scale and are used in our peripheral vision as well as to avoid
stubbing our toes in twilight conditions when we can’t really see in color.
Our cones detect fine detail and color.
Concentrated near the retina’s central focal point called the fovea, cones function only
in well lit conditions, allowing you to appreciate the intricacies of your grandma’s china
pattern or your uncle’s sleeve tattoo.
And the human eye is terrific at seeing color.
Our difference threshold for colors is so exceptional that the average person can distinguish
a million different hues.
There’s a good deal of ongoing research around exactly how our color vision works.
But two theories help us explain some of what we know.
One model, called the Young-Helmholtz trichromatic theory suggests that the retina houses three
specific color receptor cones that register red, green and blue, and when stimulated together,
their combined power allows the eye to register any color.
Unless, of course you’re colorblind.
About one in fifty people have some level of color vision deficiency.
They’re mostly dudes because the genetic defect is sex linked.
If you can’t see the Crash Course logo pop out at you in this figure, it’s likely that
your red or green cones are missing or malfunctioning which means you have dichromatic instead of
trichromatic vision and can’t distinguish between shades of red and green.
The other model for color vision, known as the opponent-process theory, suggests that
we see color through processes that actually work against each other.
So some receptor cells might be stimulated by red but inhibited by green, while others
do the opposite, and those combinations allow us to register colors.
But back to your eyeballs.
When stimulated, the rods and cones trigger chemical changes that spark neural signals
which in turn activate the cells behind them called bipolar cells, whose job it is to turn
on the neighboring ganglion cells.
The long axon tails of these ganglions braid together to form the ropy optic nerve, which
is what carries the neural impulses from the eyeball to the brain.
That visual information then slips through a chain of progressively complex levels as
it travels from optic nerve, to the thalamus, and on to the brain’s visual cortex.
The visual cortex sits at the back of the brain in the occipital lobe, where the right
cortex processes input from the left eye and vice versa.
This cortex has specialized nerve cells, called feature detectors that respond to specific
features like shapes, angles and movements.
In other words different parts of your visual cortex are responsible for identifying different
aspects of things.
A person who can’t recognize human faces may have no trouble picking out their set
of keys from a pile on the counter.
That’s because the brains object perception occurs in a different place from its face
perception.
In the case of Dr. Sacks, his condition affects the region of the brain called the fusiform
gyrus, which activates in response to seeing faces.
Sacks’s face blindness is congenital, but it may also be acquired through disease or
injury to that same region of the brain.
And some cells in a region may respond to just one type of stimulus, like posture or
movement or facial expression, while other clusters of cells weave all that separate
information together in an instant analysis of a situation.
That clown is frowning and running at me with a tiny cream pie.
I’m putting these factors together.
Maybe I should get out of here.
This ability to process and analyze many separate aspects of the situation at once is called
parallel processing.
In the case of visual processing, this means that the brain simultaneously works on making
sense of form, depth, motion and color and this is where we enter the whole world of
perception which gets complicated quickly, and can even get downright philosophical.
So we’ll be exploring that in depth next time but for now, if you were paying attention,
you learned the difference between sensation and perception, the different thresholds that
limit our senses, and some of the neurology and biology and psychology of human vision.
Thanks for watching this lesson with your eyeballs, and thanks to our generous co-sponsors
who made this episode possible: Alberto Costa, Alpna Agrawal PhD, Frank Zegler, Philipp Dettmer
and Kurzgesagt.
And if you’d like to sponsor an episode and get your own shout out, you can learn
about that and other perks available to our Subbable subscribers, just go to subbable.com/crashcourse.
This episode was written by Kathleen Yale, edited by Blake de Pastino, and our consultant
is Dr. Ranjit Bhagwat.
Our director and editor is Nicholas Jenkins, the script supervisor is Michael Aranda who
is also our sound designer, and our graphics team is Thought Cafe.
Click on any text or timestamp to jump to that moment in the video
Share:
Most transcripts ready in under 5 seconds
One-Click Copy125+ LanguagesSearch ContentJump to Timestamps
Paste YouTube URL
Enter any YouTube video link to get the full transcript
Transcript Extraction Form
Most transcripts ready in under 5 seconds
Get Our Chrome Extension
Get transcripts instantly without leaving YouTube. Install our Chrome extension for one-click access to any video's transcript directly on the watch page.