HELP WITH PSYCHOLOGY CLASS

posted 

Save Time On Research and Writing
Hire a Pro to Write You a 100% Plagiarism-Free Paper.
Get My Paper

Prepare answers to each of the questions provided. Each answer should be approximately 1 page long, double-spaced, but feel free to go beyond that (up to two pages). Provide original examples to help clarify definitions or concepts. If you use outside references (which you should) include them on a separate reference sheet (as well as reference them in your answers).

Grading

You will be marked on the following:

· accuracy of your answer

Save Time On Research and Writing
Hire a Pro to Write You a 100% Plagiarism-Free Paper.
Get My Paper

· completeness of your answer

· the degree to which your answer is original (does not quote the text; adds examples)

Each question will be out of 10 points. Here is what certain scores mean:

6/10

Most basic points are there but no examples or applications or thoughts beyond what is directly in the text. Some minor points may be missing.

7/10

All points, major and minor, are included in the answer.

8/10

All points are included in the answer and there is some original contribution.

9-10/10

Not only is the question answered but the student also makes the effort to take it beyond the text and ties in materials from elsewhere

Questions

1. The book mentions the classic idea of there being only five senses, but also shows that this idea is false. Table 1.1 lists 12 total senses instead. So, why do you think the popular idea of five senses has persisted? In thinking about this, consider on what basis does one define a sense? Following this line of thought, can you list any sense not included in Table 1.1?

2. When using the Method of Limits to measure a threshold, the crossover point between detecting vs. not-detecting a threshold may be different for ascending and descending trials. Why do you think this happens? Researchers usually average together the results of ascending and descending trials. Do you think this is valid?

3. Does it make sense to separately define the absolute threshold and difference threshold? Can you see how they might be the same?

4. Traditionally, the bridges of submarines were illuminated with red light, especially during nighttime operations. Given your knowledge of scotopic vision, why do you think this was done?

5. When vision loss occurs in adulthood, there are various ways to restore function. However, restoration does not require that patients learn how to see all over again, as noted in the textbook’s description of the Argus II device. However, in some parts of the world, children born with cataracts may not have access to treatment until they are teenagers, or older. These children do have to learn how to see. What do you think accounts for this difference in response to treatment?

6. Imagine seeing the letter R. A top-down approach to letter recognition would suggest that you have some knowledge of what the letter R should look like, and this knowledge is used to enable recognition. But think about all of the variations in the letter R – not just in computer fonts but also the endless variety of human handwriting. How does our knowledge of what an R should look like compare to the variety of letters we encounter? How does this knowledge enable recognition when the seen letter is not exactly the same as the mental representation of R? Is it possible to store every possible variation? How can this problem be solved? Include in your answer some (at least four) of the more extreme forms of the letter “R” that can still be recognized as the letter “R” in the proper context.

7. In Chapter 3, the Purkinje shift was discussed in relation to the transition from photopic (cone) vision to scotopic (rod) vision during dark adaptation. This phenomenon is associated with, in dim illumination, objects reflecting long wavelength light to appear darker than objects reflecting short wavelength light. Using the information presented in Figure 6.13, can you explain why this is so?

8. The most recent evolutionary change to colour vision was the L-cone system, completing the trichromatic system by adding the ability to discriminate the longer wavelengths from the short and middle wavelengths. Further, trichromacy occurs almost exclusively in primates. What do you think might be the evolutionary advantage of this adaptation?

See the Assignment Schedule for the due date.
Submit using the Dropbox.

Praise for
“This book gives students the advantage of
having access to a lot of resources and
ENGAGING ACTIVITIES that will help them
understand the concepts as they read each
chapter in the text.”
—Nicha K. Otero, University of Arkansas—Fort Smith
“This is an EXCELLENT TEXTBOOK for an
undergraduate class in sensation and
perception. The textbook covers all the
important topics in the field. The authors
include a good mix of classic research and
recent research. The chapters are well
written and well organized. Students will
enjoy learning from this book.”
—Robert Hines, University of Arkansas, Little Rock
“The CLARITY OF WRITING makes this product
very reader friendly.”
—T. C. Sim, Sam Houston State University
“The application sections are interesting and
allow students to see how what they are
learning is RELEVANT TO THE REAL WORLD.”
—Alexis Grosofsky, Beloit College
“This text is EASY TO READ, includes examples
that are relevant to your life, covers all the
essential information for a course in
sensation and perception.”
—Jennifer L. Thomson, Messiah College
“The INTERACTIVE NATURE of these [ISLE]
activities is great for student learning. The
function of emailing themselves their action
plan is also useful for student organization.
Flashcards are great for revision and
checking on knowledge. I think these
resources will be used a lot by the students.”
—Sue Wilkinson, Cardiff Metropolitan University

Sara Miller McCune founded SAGE Publishing in 1965 to support
the dissemination of usable knowledge and educate a global
community. SAGE publishes more than 1000 journals and over
800 new books each year, spanning a wide range of subject areas.
Our growing selection of library products includes archives, data,
case studies and video. SAGE remains majority owned by our
founder and after her lifetime will become owned by a charitable
trust that secures the company’s continued independence.
Los Angeles | London | New Delhi | Singapore | Washington DC | Melbourne

Florida International University Hanover College

Copyright © 2019 by SAGE Publications, Inc.
All rights reserved. No part of this book may be reproduced or utilized in any form or
by any means, electronic or mechanical, including photocopying, recording, or by
any information storage and retrieval system, without permission in writing from the
publisher.
Illustrations provided by Auburn Associates and Body Scientific International, LLC.
Special thanks to Carolina Hrejsa, Senior Medical Illustrator of Body Scientific
International.
Printed in the United States of America
Library of Congress Cataloging-in-Publication Data
Names: Schwartz, Bennett L., author. | Krantz, John H., author.
Title: Sensation and perception / Bennett L. Schwartz, Florida International University,
John H. Krantz, Hanover College.
Description: Los Angeles : SAGE, [2019] | Includes bibliographical references and index.
Identifiers: LCCN 2017040166 | ISBN 9781506383910 (hardcover : alk. paper)
Subjects: LCSH: Senses and sensation. | Perception.
Classification: LCC QP430 .S334 2019 | DDC 612.8—dc23
LC record available at https://lccn.loc.gov/2017040166
This book is printed on acid-free paper.
18 19 20 21 22 10 9 8 7 6 5 4 3 2 1
FOR INFORMATION:
SAGE Publications, Inc.
2455 Teller Road
Thousand Oaks, California 91320
E-mail: order@sagepub.com
SAGE Publications Ltd.
1 Oliver’s Yard
55 City Road
London EC1Y 1SP
United Kingdom
SAGE Publications India Pvt. Ltd.
B 1/I 1 Mohan Cooperative Industrial Area
Mathura Road, New Delhi 110 044
India
SAGE Publications Asia-Pacific Pte. Ltd.
3 Church Street
#10-04 Samsung Hub
Singapore 049483
Acquisitions Editor: Abbie Rickard
Content Development Editor: Emma Newsom
Editorial Assistant: Jennifer Cline
Production Editor: Jane Haenel
Copy Editor: Colleen Brennan
Typesetter: C&M Digitals (P) Ltd.
Proofreader: Scott Oney
Indexer: Sheila Bodell
Cover and Interior Designer: Scott Van Atta
Marketing Manager: Jenna Retana

ISLE Activities xvi
Preface xxii
Acknowledgments xxvii
About the Authors xxix
1 WHAT IS PERCEPTION? …………………………………………………1
2 RESEARCH METHODOLOGY ………………………………………….27
3 VISUAL SYSTEM: THE EYE ……………………………………………53
4 VISUAL SYSTEM: THE BRAIN ………………………………………..91
5 OBJECT PERCEPTION ………………………………………………..123
6 COLOR PERCEPTION ………………………………………………….151
7 DEPTH AND SIZE PERCEPTION …………………………………….187
8 MOVEMENT AND ACTION …………………………………………..227
9 VISUAL ATTENTION ………………………………………………….255
10 THE AUDITORY SYSTEM …………………………………………….289
11 THE AUDITORY BRAIN AND SOUND LOCALIZATION ………….317
12 SPEECH PERCEPTION ………………………………………………..341
13 MUSIC PERCEPTION ………………………………………………….371
BRIEF CONTENTS
istockphoto.com

14 TOUCH AND PAIN …………………………………………………….401
15 OLFACTION AND TASTE ……………………………………………..435
Glossary 461
References 475
Author Index 493
Subject Index 500

DETAILED CONTENTS
ISLE Activities, xvi
Preface, xxii
Acknowledgments, xxvii
About the Authors, xxix
1 WHAT IS PERCEPTION? …………………….. 1
Introduction, 1
Why Is This Psychology? 3
The Myth of Five Senses, 4
The Basics of Perception, 5
Action, 8
The Nature of Experience
and Phenomenology, 8
The History of Sensation and Perception, 9
The Beginnings, 10
Helmholtz Versus Hering, 11
Weber, Fechner, and the Birth
of Psychophysics, 12
The 20th Century and the Study
of Perception: Cognitive
Psychology Approaches, 13
Gestalt Psychology, 13
Direct Perception (The Gibsonian
Approach), 15
Information-Processing Approach, 15
Computational Approach, 16
Neuroscience in Sensation and Perception, 17
Exploration: Cognitive Penetration, 19
Application: Avoiding Collisions, 21
Chapter Summary, 23
Review Questions, 23
Ponder Further, 24
Key Terms, 24
2 RESEARCH METHODOLOGY ……………… 27
Introduction, 27
The Measures and Methods
of Psychophysics, 29
Method of Limits, 29
Method of Constant Stimuli, 32
Method of Adjustment, 33
Magnitude Estimation, 34
Catch Trials and Their Use, 35
Signal Detection Theory, 36
Neuroimaging Techniques, 42
Exploration: Intersensory
Perception, 44
Application: Psychophysics in
Assessment: Hearing Tests
and Vision Tests, 45
Chapter Summary, 49
Review Questions, 50
Ponder Further, 50
Key Terms, 50
3 VISUAL SYSTEM: THE EYE ………………. 53
Introduction, 53
Light, 55
The Eye and Its Role in
the Visual System, 57
Field of View, 58
Anatomy of the Eye, 58
The Cornea, 59
The Lens, 60
Development: The Emerging
and Aging Eye, 61
The Retina, 62
Anatomy of the Retina, 63
The Receptors: Rods
and Cones, 64
Retinal Physiology, 65
Transduction of Light, 66
Classes of Receptors, 68
The Duplex Theory of Vision, 68
Spectral Sensitivity and the
Purkinje Shift, 69
Spatial Summation and Acuity, 70

Dark and Light Adaptation, 70
Development: Infant Acuity, 72
Retinal Ganglion Cells and
Receptive Fields, 73
Refractive Errors and Diseases
of the Eye, 77
Myopia (Nearsightedness), 78
Hyperopia (Farsightedness) and
Presbyopia (Old-Sightedness), 78
Astigmatism, 79
Cataracts, 80
Macular Degeneration, 80
Retinitis Pigmentosa, 80
Exploration: Animal Eyes, 82
Cats, 82
Nautiluses, 83
Application: Vision Prostheses, 85
Chapter Summary, 86
Review Questions, 87
Ponder Further, 88
Key Terms, 88
4 VISUAL SYSTEM: THE BRAIN …………… 91
Introduction, 91
The Optic Nerve and Chiasm, 92
The Lateral Geniculate Nucleus, 94
Processing in the LGN, 97
The Superior Colliculus, 98
The Primary Visual Cortex, 99
Mapping the Eye on the Brain, 101
Receptive Fields of V1 Cells, 102
Simple Cells, 103
Complex Cells and V1 Responses
to Visual Features, 103
The Organization of V1, 104
V2 and Beyond, 106
V2, 106
Functional Pathways in the
Visual Cortex, 107
The Ventral Pathway, 109
The Dorsal Pathway, 110
Where Does Vision Come Together? 111
Development of the Visual System, 111
Exploration: Blindsight, 113
Application: Conjugate Gaze Palsy, 116
Chapter Summary, 117
Review Questions, 119
Ponder Further, 119
Key Terms, 119
5 OBJECT PERCEPTION …………………….. 123
Introduction, 123
Introduction to Object Perception, 124
Top-Down Processing and Bottom-Up
Processing, 126
Recognition and Representation, 127
Perceptual Organization, 128
Gestalt Psychology and Perceptual
Organization, 128
Figure–Ground Organization, 129
A Few Rules That Govern What
We See as Figure and What
We See as Ground, 131
Gestalt Laws of Perceptual
Grouping, 132
Perceptual Interpolation, 134
Recognition by Components, 136
The Neuroanatomy and Physiology
of Object Perception, 137
Representation of Shapes in Area V4, 137
Object Recognition in the
Inferotemporal Area, 137
The Fusiform Face Area and
Face Recognition, 138
Prosopagnosia, 139
Other Inferotemporal Cortex
Areas With Specific Object
Recognition Functions, 140
Grandmother Cells and Specific Coding
in the Inferotemporal Cortex, 140
Exploration: Vision and Animacy: How
Do We Tell a Who From a What? 143
Application: The Science of
Airport Screening, 145
Chapter Summary, 147
Review Questions, 148

Ponder Further, 148
Key Terms, 148
6 COLOR PERCEPTION …………………….. 151
Introduction, 151
Wavelengths of Light and Color, 153
Hue, Saturation, Lightness,
and Brightness, 155
Additive and Subtractive Color Mixing, 156
Additive Color Mixing (Mixing Lights), 157
Subtractive Color Mixing
(Mixing Paints), 158
Color-Matching Experiments, 158
The Retina and Color, 160
Univariance, or Why More Than
One Receptor Is Necessary
to See in Color, 161
The Trichromatic Theory of Color Vision, 162
The Opponent-Process Theory
of Color Perception, 163
Findings That Support Opponent-
Process Theory, 164
Hue Cancellation, 165
Opponent Cells in the LGN and V1, 166
The Development of Color Perception, 168
Color Perception in Infancy, 168
Aging and Color Perception, 169
Color Deficiency, 169
Rod Monochromacy, 171
Cone Monochromacy, 171
Dichromacy, 172
Protanopia, 172
Deuteranopia, 172
Tritanopia, 173
Cortical Achromatopsia, 174
Constancy: Lightness and
Color Constancy, 175
Color Constancy, 175
Lightness Constancy, 177
Exploration: The Color Purple, 178
Application: Color Deficiency and
Our Colorful Information, 180
Chapter Summary, 181
Review Questions, 182
Ponder Further, 183
Key Terms, 183
7 DEPTH AND SIZE PERCEPTION ……….. 187
Introduction, 187
Monocular Depth Cues, 189
Occlusion (or Interposition), 190
Relative Height, 191
Relative Size, 191
Familiar Size, 192
Linear Perspective, 193
Texture Gradients, 194
Atmospheric Perspective, 194
Shadows and Shading, 195
Motion Cues, 196
Motion Parallax, 196
Deletion and Accretion, 197
Optic Flow, 198
Oculomotor Cues, 199
Accommodation, 199
Vergence (or Convergence), 199
Binocular Cues to Depth, 200
Binocular Disparity, 201
Corresponding and
Noncorresponding Points, 202
The Correspondence
Problem, 206
Stereograms, 207
Random-Dot Stereograms, 208
The Anatomy and Physiology
of Binocular Perception, 209
Developmental Issues in
Stereopsis, 210
Size Perception, 211
Size Constancy, 212
Visual Illusions of Size
and Depth, 213
The Müller–Lyer Illusion, 214
The Ponzo Illusion, 215
The Ames Room Illusion, 215
The Moon Illusion, 216

Exploration: Stereopsis and Sports: Do We
Need Binocular Vision in Sports? 218
Application: Virtual Reality
and Therapy, 219
Chapter Summary, 222
Review Questions, 223
Ponder Further, 223
Key Terms, 223
8 MOVEMENT AND ACTION ……………… 227
Introduction, 227
How Do We Perceive Motion? 228
Motion Thresholds: How Slow
and How Fast? 228
Real and Apparent Motion, 230
The Neuroscience of Vision
and Motion, 232
Motion Detection in the Retina, 232
The Complexity of Motion, 232
Corollary Discharge Theory, 234
Eye Movements, 235
Saccades, 236
Smooth-Pursuit Eye Movements, 236
MT: The Movement Area of
the Brain, 236
Motion Aftereffects, 239
Form Perception and Biological
Motion, 241
Action, 242
Visually Guided Eye Movements, 244
Visually Guided Grasping, 245
Exploration: Motion Illusions, 246
Illusion 1: Rotating Snakes, 247
Illusion 2: Illusory Rotation, 248
Illusion 3: The Furrow Illusion, 248
Application: Motion Perception
in Airplane Pilots, 249
Chapter Summary, 251
Review Questions, 251
Ponder Further, 252
Key Terms, 252
9 VISUAL ATTENTION ……………………… 255
Introduction, 255
Selective Attention, 258
Spatial Limits of Attention, 259
Attention and the Direction of
Gaze in Space, 259
Inattentional Blindness, 262
Stimulus Features That Draw Attention, 264
Stimulus Salience, 265
Visual Search, 266
Feature Integration Theory, 267
Attention Over Time, 268
Change Blindness, 268
Attentional Blink and Rapid Serial
Visual Presentation, 270
The Anatomy and Physiology
of Attention, 272
The Orienting Attention Network, 273
The Executive Attention Network, 273
How Attention Affects the
Visual Brain, 274
The Neuropsychology of Attention, 275
Bálint’s Syndrome, 277
Developmental Aspects of
Visual Attention, 278
Exploration: Awareness and
Visual Consciousness, 279
Perceptual Bistability, 280
Blindsight, 281
Application: Distracted Driving, 282
Chapter Summary, 284
Review Questions, 285
Ponder Further, 285
Key Terms, 286
10 THE AUDITORY SYSTEM ……………….. 289
Introduction, 289
Sound as Stimulus, 290
The Relation of Physical and Perceptual
Attributes of Sound, 292
Amplitude and Loudness, 292

Frequency and Pitch, 294
Waveform and Timbre, 295
Phase, 297
Anatomy of the Ear, 298
The Outer Ear, 298
The Middle Ear, 299
The Inner Ear, 301
The Basilar Membrane of the Cochlea, 302
The Organ of Corti, 304
Exploration: Hearing Loss, 306
Conductive Hearing Loss, 307
Sensorineural Hearing Loss, 307
Tinnitus, 308
Application: The Science of Hearing
Aids and Cochlear Implants, 308
Hearing Aids, 308
Cochlear Implants, 310
Chapter Summary, 311
Review Questions, 312
Ponder Further, 313
Key Terms, 313
11 THE AUDITORY BRAIN AND SOUND
LOCALIZATION …………………………..317
Introduction, 317
Brain Anatomy and the Pathway
of Hearing, 318
Auditory Nerve Fibers, 318
Auditory Cortex, 320
Localizing Sound, 321
Interaural Time Difference, 322
Interaural Level Difference, 324
The Cone of Confusion, 324
Elevation Perception, 325
Detecting Distance, 325
Auditory Scene Analysis, 326
Temporal Segregation, 328
Spatial Segregation, 329
Spectral Segregation, 329
Auditory Development, 330
Exploration: Biosonar in Bats
and Dolphins, 331
Application: Concert Hall
Acoustics and Hearing, 335
Chapter Summary, 336
Review Questions, 337
Ponder Further, 337
Key Terms, 338
12 SPEECH PERCEPTION ……………………. 341
Introduction, 341
The Human Voice as Stimulus, 342
Vowels and Consonants, 343
Speech, 344
Variability in the Acoustics of Phonemes, 347
Coarticulation, 348
Categorical Perception, 349
The Effect of Vision on Speech Perception
and the McGurk Effect, 351
Top-Down Processing and
Speech Perception, 352
The Phonemic Restoration Effect, 353
Theories of Speech Perception, 355
The Development of Phoneme
Perception, 357
Speech Perception and the Brain, 358
Exploration: Hearing Loss and
Speech Perception, 362
Application: Hey Siri, or Are You Cortana?
Computer Speech Recognition, 364
Chapter Summary, 366
Review Questions, 367
Ponder Further, 367
Key Terms, 367
13 MUSIC PERCEPTION …………………….. 371
Introduction, 371
The Acoustics of Music, 373
Pitch, Chroma, and the Octave, 373
The Octave, 375

Consonance and Dissonance, 376
Dynamics and Rhythm, 377
Timbre, 378
Melody, 380
Scales and Keys and Their
Relation to Melody, 380
Gestalt Principles of Melody, 381
The Neuroscience of Music, 383
The Neuroanatomy of Music, 384
Synesthesia, 385
The Neuropsychology of Music, 386
Learning, Culture, and Music
Perception, 388
Music and Language, 388
Culture and Music Perception, 389
Exploration: Musical Illusions, 391
Shepard Tones, 392
The Octave Illusion, 392
The Scale Illusion, 393
The Tritone Paradox, 394
Application: Music Perception in
Hearing-Impaired Listeners, 394
Chapter Summary, 396
Review Questions, 397
Ponder Further, 398
Key Terms, 398
14 TOUCH AND PAIN ………………………… 401
Introduction, 401
The Skin and Its Receptors, 403
Mechanoreception, 404
SAI Mechanoreceptors, 405
SAII Mechanoreceptors, 405
FAI Mechanoreceptors, 405
FAII Mechanoreceptors, 406
Proprioception: Perceiving
Limb Position, 406
Thermoreception, 407
Nociception and the
Perception of Pain, 409
Neural Pathways, 410
Somatosensory Cortex, 413
Suborganization of the
Somatosensory Cortex, 414
Pathways for Pain, 415
The Neurochemistry of Pain:
Endogenous Opioids, 417
The Perception of Itch, 418
Haptic Perception, 419
Reading Braille, 420
Tactile Agnosia, 421
The Development of Haptic
Perception, 422
The Vestibular System: The
Perception of Balance, 423
Exploration: Electroreception in Fish, 425
Application: Phantom Limbs and
Phantom Limb Pain, 427
Chapter Summary, 429
Review Questions, 430
Ponder Further, 431
Key Terms, 431
15 OLFACTION AND TASTE ………………… 435
Introduction, 435
Olfaction, 436
The Anatomy and Physiology of
the Olfactory System, 437
The Nose, 437
Genes and Olfaction, 439
The Trigeminal Nerve, 440
The Pathway to the Brain, 440
Representation Within the
Piriform Cortex, 442
The Orbitofrontal Cortex, 443
Olfactory Perception, 443
Detection, 443
Identifying Odors, 444
Odor Imagery, 445
Olfactory Illusions, 445
Taste Perception, 447
Anatomy of the Tongue and
Taste Coding, 448

Taste and Flavor, 451
Individual Differences in Taste
Perception, 451
The Wonderful World of
Chili Peppers, 452
Development of Taste
Perception, 453
Exploration: Anosmia, 453
Application: Artificial Sweeteners
and Taste Perception, 455
Chapter Summary, 457
Review Questions, 458
Ponder Further, 459
Key Terms, 459
Glossary, 461
References, 475
Author Index, 493
Subject Index, 500

xvi Sensation and Perception
ISLE ACTIVITIES
ISLE 1.1 Sequence of Sensory Events
ISLE 1.2 Illustration of Action
ISLE 1.3 Bach’s Violin Partita No. 2 in D minor
ISLE 1.4 Column Taper Illusion
ISLE 1.5 Aristotle’s Illusion
ISLE 1.6 Motion Aftereffect
Chapter 1
ISLE 1.7 Fechner’s Colors and Benham’s Top
ISLE 1.8 Kanizsa Triangle
ISLE 1.9 Optic Flow
ISLE 1.10 Kuffler and Single-Cell Recording
ISLE 1.11 Size–Arrival Effect
Preface
ISLE P.1 Obtaining or Creating Anaglyph Glasses
ISLE 2.1
North Carolina Hot Sauce Contest
Link | Proceed to Quiz
ISLE 2.2 Method of Limits
ISLE 2.2a Dot Threshold
ISLE 2.2b Frequency Discrimination
ISLE 2.3 Method of Constant Stimuli
ISLE 2.3a Dot Threshold
ISLE 2.3b Frequency Discrimination
ISLE 2.4 Method of Adjustment
ISLE 2.4a Dot Threshold
ISLE 2.4b Frequency Discrimination
ISLE 2.5 Point of Subjective Equality
ISLE 2.6 Magnitude Estimation
ISLE 2.6a Dot Brightness
ISLE 2.6b Tone Loudness
ISLE 2.6c Line Length
Chapter 2
ISLE 2.7 Stevens’s Power Law
ISLE 2.8 Forced-Choice Method
ISLE 2.8a Dot Brightness
ISLE 2.8b Frequency Discrimination
ISLE 2.9 Signal Detection Experiment
ISLE 2.10 Signal Detection Theory
ISLE 2.10a
Basics of Signal
Detection Theory
ISLE 2.10b Visualizing Signal Detection Theory
ISLE 2.10c
Decision Making in Signal
Detection Theory
ISLE 2.11
Signal Detection Theory and the
Receiver-Operating Characteristic
(ROC) Curve
ISLE 2.12 Masking Demonstration
ISLE 2.13
Seeing With Myopia and
Presbyopia

Experiment Interactive Link Video Audio illustration

ISLE Activities xvii
ISLE 3.1 The Basics of Waves
ISLE 3.2 Accommodation
ISLE 3.3 Presbyopia
ISLE 3.4 Letters and the Fovea
ISLE 3.5 Map Your Blind Spot
ISLE 3.6 Photopic vs. Scotopic Vision
ISLE 3.7 Purkinje Shift
ISLE 3.8 Convergence
ISLE 3.9 Dark Adaptation Function
ISLE 3.9a Illustrating Dark Adaptation
ISLE 3.10
Automatic Light Adjustment in
Cockpits
Chapter 3
ISLE 3.11 Center-Surround Receptive Fields
ISLE 3.11a Simulating Kuffler’s Experiment
ISLE 3.11b
Center-Surround Receptive Fields
as Contrast Detectors
ISLE 3.12 Mach Bands
ISLE 3.13 Correcting Myopia and Hyperopia
ISLE 3.14 Astigmatism
ISLE 3.15
Living With Macular Degeneration
and Retinitis Pigmentosa
ISLE 3.16 Compound Eyes
ISLE 3.17 Vision Prostheses
ISLE 4.1 From Eye to LGN
ISLE 4.2 Simple Cells
ISLE 4.3 Complex Cells
Chapter 4
ISLE 4.4 Hypercolumns
ISLE 4.5 Navigation in Blindsight
ISLE 5.1 Segregation and Grouping
ISLE 5.2
Ambiguous Figure–Ground
Perception
ISLE 5.3 Figure–Ground Symmetry
ISLE 5.4 Gestalt Laws
ISLE 5.4a Law of Good Continuation
ISLE 5.4b Laws of Proximity and Similarity
ISLE 5.4c Law of Symmetry
Chapter 5
ISLE 5.4d Law of Closure
ISLE 5.4e Law of Common Fate
ISLE 5.5 Necker Cube
ISLE 5.6 Illusory Contours
ISLE 5.7 Geons
ISLE 5.8
Facial Responses in the Brain
(Inferotemporal Region)
ISLE 6.1 Different Types of White Lights
ISLE 6.2 Dimensions of Color
ISLE 6.3 Newton’s Prism Experiment
Chapter 6
ISLE 6.4 Color Mixing
ISLE 6.4a Additive Color Mixing
ISLE 6.4b Subtractive Color Mixing
ISLE 6.4c
Additive and Subtractive Color
Mixing

xviii Sensation and Perception
ISLE 6.5
Color-Matching Experiment:
Metameric Matches
ISLE 6.6
Trichromatic Theory and Cone
Responses
ISLE 6.7
Univariance and Color Matching
in Monochromat or During
Scotopic Vision
ISLE 6.8 Color Aftereffect
ISLE 6.8a
Color Aftereffect Using
Photographs
ISLE 6.9 Simultaneous Color Contrast
ISLE 6.10 Hue Cancellation
ISLE 6.11
Single- and Double-Opponent
Cells
ISLE 6.12 Color-Deficiency Tests
ISLE 6.13 Rod Monochromat Vision
ISLE 6.14 Dichromacy
ISLE 6.14a Cones Missing in Dichromacy
ISLE 6.14b Color Matching in Dichromats
ISLE 6.14c Simulating Dichromacy
ISLE 6.15 Illumination and Color Constancy
ISLE 6.16 Lightness Constancy
ISLE 6.17 Gelb Effect
ISLE 6.18
Color Camouflage and
Dichromacy
ISLE 7.1
Depth Perception and Adanson’s
Jumping Spider
ISLE 7.2 Monocular Depth Cues
ISLE 7.2a Relative Size
ISLE 7.2b Size Constancy
ISLE 7.2b.1 Size Constancy Illustration
ISLE 7.2b.2 Size Constancy Experiment
ISLE 7.3 Motion Depth Cues
ISLE 7.3a Motion Parallax
ISLE 7.3a1 Motion Parallax Illustrated
ISLE 7.3a2 Motion Parallax Explained
ISLE 7.3b Deletion and Accretion
ISLE 7.3c Optic Flow
ISLE 7.4 Accommodation
ISLE 7.5 Vergence
ISLE 7.6 Stereopsis
Chapter 7
ISLE 7.7 Binocular Disparity
ISLE 7.8
The Construction of Visual Depth With
Binocular Disparity
ISLE 7.8a The Horopter
ISLE 7.8b Panum’s area of fusion
ISLE 7.8c Crossed vs. Uncrossed Disparities
ISLE 7.9 Anaglyph Stereograms
ISLE 7.9a
Stereo Jitter (Does not Need
Glasses)
ISLE 7.10 Random-Dot Stereograms
ISLE 7.11 Retinal Image Size and Distance
ISLE 7.12 Müller–Lyer Illusion
ISLE 7.13 Ponzo Illusion
ISLE 7.14 Ames Room
ISLE 7.15 Moon Illusion
ISLE 7.16 Virtual Reality Photographs
ISLE 7.17 Virtual Reality and Therapy
ISLE 8.1 Relative Motion
ISLE 8.1a
Relative Motion and Frame of
Reference
ISLE 8.2 Time-Lapse Motion
ISLE 8.3 Motion Thresholds
Chapter 8
ISLE 8.4 Apparent Motion
ISLE 8.5
Correspondence Problem in Motion
(Wagon Wheel Effect)
ISLE 8.6 Induced Motion
ISLE 8.7 Reichardt Detectors

xix ISLE Activities
ISLE 8.8 Corollary Discharge
ISLE 8.8a
Corrollary Discharge
Demonstration
ISLE 8.8b Moving Afterimages
ISLE 8.9 Eye Movements
ISLE 8.10 Correlated Motion
ISLE 8.11 Motion Aftereffect
ISLE 8.11a Waterfall Illusion
ISLE 8.11b Spiral Aftereffect
ISLE 8.12 Structure From Motion
ISLE 8.13 Biological Motion
ISLE 8.13a Examples of it Working
ISLE 8.13b
Examples of the Breakdown of
Biological Motion
ISLE 8.14 Cat and Laser Pointer
ISLE 8.15 Optic Flow
ISLE 8.16 Rotating Snakes
ISLE 8.17 Spiral Staircase
ISLE 8.18 Illusory Rotation
ISLE 8.19 The Furrow Illusion
ISLE 9.1 The Stroop Effect
ISLE 9.2 Scanning
ISLE 9.3 Spatial Cuing
ISLE 9.4 Inattentional Blindness Examples
ISLE 9.5 Stimulus Salience
ISLE 9.6 Feature vs. Conjunction Search
ISLE 9.7 Change Blindness
ISLE 9.8 RSVP
ISLE 9.9
Attentional Blink and Repetition
Blindness
ISLE 9.9a
Attentional Blink and Repetition
Blindness : Demonstration
ISLE 9.9b
Attentional Blink and Repetition
Blindness : Experiment
Chapter 9
ISLE 9.10 Hemifield or Unilateral Neglect
ISLE 9.11 Bálint’s Syndrome
ISLE 9.12 Perceptual Bistability
ISLE 9.12a Necker Cube
ISLE 9.12b Vase/Face
ISLE 9.13 Navigation in Blindsight
ISLE 10.1 The Sound Stimulus
ISLE 10.1a Sound and Air Pressure
ISLE 10.1b The Sound Wave and Distance
ISLE 10.1c Pure Tones
ISLE 10.2
The Speed of Sound and the
Sonic Boom
ISLE 10.3 The Decibel Scale
Chapter 10
ISLE 10.4 Frequency and Pitch on a Piano
ISLE 10.5 Frequency Response of the Ear
ISLE 10.6 Timbre and Musical Instruments
ISLE 10.7 Fourier Analysis in Audition
ISLE 10.8 Missing Fundamental
ISLE 10.9 Timbre and Overtones

xx Sensation and Perception
ISLE 11.1 Sound Localization Experiment
ISLE 11.2 Interaural Time Differences
ISLE 11.2a
Auditory Illustration (Needs
Headphones)
ISLE 11.2b Interactive Diagram
ISLE 11.3 Interaural Level Differences
ISLE 11.3a
Auditory Illustration (Needs
Headphones)
ISLE 11.3b Interactive Diagram
Chapter 11
ISLE 11.4 Auditory Scene Analysis
ISLE 11.5 Doppler Shift
ISLE 11.6 Architectural Space and Echoes
ISLE 11.7 Testing a Concert Hall
ISLE 12.1 Coarticulation
ISLE 12.2 Voicing-Onset Time
ISLE 12.3 McGurk Effect
ISLE 12.4 Familiar vs. Unfamiliar Languages
Chapter 12
ISLE 12.5 Phonemic Restoration Effect
ISLE 12.6 Genie
ISLE 12.7 Broca’s and Wernicke’s Aphasia
ISLE 13.1 Kurdish Music Example
ISLE 13.2
Javanese Gamelan Music
Example
ISLE 13.3 Ancient Greek Music
ISLE 13.4 35,000-Year-Old Flute
ISLE 13.5 Is This Music?
ISLE 13.6 The Octave and Tone Similarity
ISLE 13.7 Pentatonic Music
ISLE 13.8 Meter and Beat
ISLE 13.9 Bolero Clip
Chapter 13
ISLE 13.10 Attack and Decay
ISLE 13.11 Examples of Melody
ISLE 13.12 Types of Scales
ISLE 13.12a Play Scales
ISLE 13.12b Natural vs. Equal Tempered
ISLE 13.13 Gestalt Principles Review
ISLE 13.13a Proximity and Similarity
ISLE 13.13b Closure
ISLE 13.13c Good Continuation
ISLE 13.14
Gestalt Principle: Proximity:
Bach’s Partita No. 3 in E major
ISLE 10.10 Phase and Cancellation
ISLE 10.11 The Middle Ear
ISLE 10.12
The Basilar Membrane and Sound
Stimuli
ISLE 10.13 The Traveling Wave
ISLE 10.14 Place Code Theory
ISLE 10.15
The Basilar Membrane and
Fourier Analysis
ISLE 10.16 Transduction and Hair Cells
ISLE 10.17 Temporal Code Theory

xxi ISLE Activities
ISLE 14.1 Action of Mechanoreceptors
ISLE 14.2
Mechanoreceptors and Aristotle’s
Illusion
ISLE 14.3 Heat Grille
ISLE 14.4 Somatosensory Pathways
Chapter 14
ISLE 14.5
Melzack and Wall’s Gate Control
Theory
ISLE 14.6
Professor Ramachandran and
Phantom Limb Syndrome
ISLE 14.7 Treatment of Phantom Limb Pain
ISLE 15.1 Brain Area for Olfactory Bulbs
Chapter 15
ISLE 15.2 Posterior Piriform Cortex
ISLE 13.15 A Shave and a Haircut
ISLE 13.16
Cross-Modal Matchings as a
Simulation of Synesthesia
ISLE 13.17 Indian Raga Music Example
ISLE 13.18
Bach’s Violin Partita No. 2 in D
minor
ISLE 13.19 Shepard Tones
ISLE 13.20 Octave Illusion
ISLE 13.21 Scale Illusion
ISLE 13.22 Tritone Paradox
Bonus Illustrations
Shifting Melody: Ratio or Constant
Musical Stroop: A Part of Cognition
Laboratory Experiments

PREFACE
Everything we do as human beings starts with our ability to perceive the world. We
wake up in the morning to the jarring sound of an alarm clock prompting us to turn
on the light in the room so we can see. A few minutes later, the aroma of the coffee
lifts our mood and gets us ready to face the day. We feel inside our pocketbooks and
briefcases to make sure we have our keys. Then we listen to the radio as we watch for
traffic on the busy streets in our cars. Thus, nearly everything we do over the course of
our waking hours involves sensation and perception. Think of the obstacles faced by
a person missing one or more of these sensory systems. The famous Helen Keller was
completely dependent on her sense of touch in the absence of vision and hearing. For
this reason, the study of sensation and perception has been important to psychologists
from the very beginnings of psychology.
Sensation and Perception is designed to be a primary textbook for undergraduate
courses in sensation and perception or, as in some universities, simply perception. We
have written it primarily with students in mind but also so that professors will find it
useful as a supplement and guide to their teaching. The two authors come from very
different settings: One is a professor at a large public university, the other at a small
private college. We have tried to meld our experiences into a textbook that will be
useful in both the large lecture typical of the public university and the small seminar
of the private college.
When we teach sensation and perception, one of the most common questions we
receive from students is the following: How is this material relevant to my life and
to my future as a psychologist? Sensation and perception textbooks seldom address
this topic. Students learn about theories, physiology, anatomy, and experiments, and
they get to see some curious illusions. These are all important to the understand-
ing of sensation and perception, and these issues are well covered in our book. But
Sensation and Perception also instructs students in how to apply these concepts to
their everyday lives and how sensation and perception research is valuable in the real
world. For example, we discuss applications of sensation and perception research
to driving cars, to playing sports, and to evaluating risk in the military. Moreover,
we discuss numerous medical applications, including extended discussion of such
topics as macular degeneration, retinitis pigmentosa, color blindness, hearing aids
and cochlear implants, and neuropsychological conditions such as prosopagnosia.
For each chapter, we finish with Exploration and Application sections. The Application
sections deal directly with applications of sensation and perception research.
The higher education classroom has changed dramatically in the past decade.
Many students take some, if not all, of their classes online without ever having to
come to a university building. Classes now usually have companion websites, and
students have been “googling” interesting topics since they first could read. A mod-
ern textbook needs to adjust to the easy availability of information. What a mod-
ern textbook offers that distinguishes it from online sources of information is its
accuracy at presenting the most relevant information in an engaging, well-written
manner. Of course, we wrote our textbook to be approachable and readable, but
fundamentally what we offer our students is the guarantee that what they read here
is an accurate reflection of what is relevant in the field of sensation and perception.
Having said that, Sensation and Perception is written with the current state of
technology in mind. First and foremost, we have prepared the Interactive Sensation
Laboratory Exercises (ISLE) website, the companion website for the textbook. The
ISLE website provides opportunities to view and hear many of the phenomena we dis-
cuss in the book, including anatomical diagrams, visual illusions, auditory illusions,

xxiii Preface
and musical selections. It provides simulations of experiments and neurological pro-
cesses, and it provides exercises to help students understand how models such as
signal detection actually work. A book cannot present sound or motion, important
components of sensation and perception. But a click on an ISLE link will bridge the
gap, bringing these phenomena to your attention. Both teacher and student will find
ISLE an invaluable resource.
We also hope that student and professor alike will find our pedagogical resources
to be helpful. To help students learn, we start each chapter with a list of learning
objectives, which should help the reader organize the material and therefore learn
it better. After each major section, we have “test your knowledge” questions. These
open-ended questions promote thinking about the material broadly and linking con-
cepts together. Key terms with must-be-learned definitions are clearly marked off
in the margins, and illustrations throughout the text reinforce what is written in
the text. At the end of each chapter is a carefully organized chapter summary, a set
of review questions, “ponder further” questions for those who want to delve even
deeper into the topic, and a list of the key terms presented in the chapter.
Sensation and Perception emphasizes science. It describes experiments, patients
with neurological disorders, the areas of the brain involved in sensation and per-
ception, and the theory that links this research together. We wrote this book with
students in mind—their concerns, interests, and curiosity. This book not only empha-
sizes the science of sensation and perception but also tells a story about our search to
understand our own minds and how we can benefit from that understanding.
WHAT IS NEW IN THIS EDITION
So why a new edition? There are several reasons: We have included new science that
has emerged, we have learned from the feedback of reviewers and users, and we have
drawn on our own experience in the classroom. We have reviewed, revised, and updated
each and every chapter. Every chapter has been thoroughly reviewed and edited. This
process has been helped extensively by several rounds of reviews on each chapter to
which we are indebted to the reviewers who have so kindly lent their critique. As we
have gone through the chapters and feedback, we have made several changes that need
to be explicitly mentioned.
We have expanded the coverage of perceptual development in this text. In addi-
tion, we have made sure that our developmental topics always are clearly identified
in the chapter headings to make sure that that material can be found. So, in almost
every chapter, there is a section titled “Development.” We also added some new sec-
tions to chapters that did not have developmental sections. For example, we added
a section on infant acuity in Chapter 3; a section on the development of color per-
ception, including both infants and aging adults, in Chapter 6; and a section on the
development of haptic perception in Chapter 14.
In the first edition, we had an In Depth section that either examined a research
topic of interest or an application. In this edition, we have created two sections to
replace the In Depth section. The Exploration section examines a current research
topic, and the Application section discusses one of the many applications of sensa-
tion and perception. In some chapters, the former In Depth section has become an
Exploration and in other chapters an Application section, and in one case, Chapter 10,
the material was divided to make both sections and to limit both sections to a more
feasible length. Table P.1 lists these sections with their titles. There were many rea-
sons for this change, but perhaps the most compelling was the fact that research
in this field is ongoing in many interesting and exciting ways, and the Exploration

xxiv Sensation and Perception
section allows us to give a taste of this excitement. It could be argued that there are
more applications of sensation and perception than any other field of psychology; the
Application section allows us to help students see a few of the ways that they interact
with this field every day.
In addition to the content changes, we have expanded and refined the pedagogi-
cal features of this text. The learning objectives have been revised to make sure that
each major heading has a learning objective, and they have been reworked to take
into account Bloom’s taxonomy. At the end of each major section, we have increased
the number of Test Your Knowledge questions. Finally, at the end of each chapter are
Ponder Further questions, which ask students to take the material of the chapter and
consider the implications beyond the confines of the material presented. In addition,
the video links in the text have all been converted to ISLE activities to make is eas-
ier for students to access these illustrations. Beyond that, there are many new ISLE
activities that enhance the coverage of existing material and cover the new material
as well.
TABLE P.1 Exploration and Application Sections at the End of Each Chapter
Chapter Exploration Application
1 Cognitive Penetration Avoiding Collisions
2 Intersensory Perception Psychophysics in Assessment:
Hearing Tests and Vision Tests
3 Animal Eyes Vision Prostheses
4 Blindsight Conjugate Gaze Palsy
5 Vision and Animacy: How Do We
Tell a Who From a What?
The Science of Airport Screening
6 The Color Purple Color Deficiency and Our Colorful
Information
7 Stereopsis and Sports: Do We
Need Binocular Vision in Sports?
Virtual Reality and Therapy
8 Motion Illusions Motion Perception in Airplane
Pilots
9 Awareness and Visual
Consciousness
Distracted Driving
10 Hearing Loss The Science of Hearing Aids and
Cochlear Implants
11 Biosonar in Bats and Dolphins Concert Hall Acoustics and
Hearing
12 Hearing Loss and Speech
Perception
Hey Siri, or Are You Cortana?
Computer Speech Recognition
13 Musical Illusions Music Perception in Hearing-
Impaired Listeners
14 Electroreception in Fish Phantom Limbs and Phantom
Limb Pain
15 Anosmia Artificial Sweeteners and Taste
Perception

xxv Preface
ANAGLYPH GLASSES
Some of the photographic figures in the text and on ISLE are anaglyphic stereograms,
and they will require special glasses to be seen properly. (See Chapter 7 for more infor-
mation about stereograms.) We provide these glasses with this edition of the text.
However, if you lose them or buy a used copy of the text without these glasses, they are
easy to make, or, if you prefer, cheap to buy, individually or in bulk. Please note that you
will need red/cyan anaglyph glasses. Consult ISLE P.1 for specific information about
either making or ordering these glasses.
DIGITAL RESOURCES
SAGE edge offers a robust online environment featuring an impressive array of
tools and resources for review, study, and further exploration, keeping both instructors
and students on the cutting edge of teaching and learning. Go to edge.sagepub.com/
schwartz2e to access the companion site.
SAGE edge for Instructors
SAGE edge for Instructors, a password-protected instructor resource site, supports
teaching by making it easy to integrate quality content and create a rich learning envi-
ronment for students. The following chapter-specific assets are available on the teach-
ing site:
• Author-created test banks provide a diverse range of questions as well as the
opportunity to edit any question and/or insert personalized questions to effec-
tively assess students’ progress and understanding.
• Lecture notes summarize key concepts by chapter to assist in the preparation of
lectures and class discussions.
• Sample course syllabi for semester and quarter courses provide suggested models
for structuring a course.
• Editable, chapter-specific PowerPoint slides offer complete flexibility for creating
a multimedia presentation for the course.
• Lively and stimulating ideas for class assignments can be used in class to reinforce
active learning. The creative assignments apply to individual or group projects.
• Chapter-specific discussion questions help launch classroom interaction by
prompting students to engage with the material and by reinforcing important
content.
• Multimedia content includes audio and video resources that appeal to students
with different learning styles.
SAGE edge for Students
SAGE edge for Students provides a personalized approach to help students accomplish
their coursework goals in an easy-to-use learning environment. The open-access study
site includes the following:
• Interactive Sensation Laboratory Exercises (ISLE) provide opportunities to view
and hear many of the phenomena discussed in the book, including anatomical
diagrams, visual illusions, auditory illusions, and musical selections.
• Learning objectives reinforce the most important material.
ISLE P.1
Obtaining or Creating
Anaglyph Glasses

xxvi Sensation and Perception
• Mobile-friendly eFlashcards strengthen understanding of key terms and concepts.
• Multimedia content includes audio and video resources that appeal to students
with different learning styles.
• EXCLUSIVE! Access to full-text SAGE journal articles that have been carefully
selected to support and expand on the concepts presented in each chapter.

ACKNOWLEDGMENTS
We need to thank many people for their contributions to this book. Although the cover
lists Bennett and John as the authors, a multitude of people made invaluable contribu-
tions to ensuring that this book came together. First, we must thank many people at
SAGE Publishing for their help with the development, art program, and production of
Sensation and Perception. We continue to thank those people who worked on the first
edition, for without their work we would not be working on a second edition. Reid
Hester initiated the project, brought us together, guided the book to the finish line, and
saw it published in a timely manner. Lucy Berbeo, Nathan Davidson, Laura Barrett
and her staff in Production, Sarita Sarak, and Jim Kelly rounded out the team for the
first edition. We gratefully acknowledge all of their work. As is so often the case, teams
change over time. We had a mostly new team for this edition, but they have really been
wonderful additions to the project. Abbie Rickard stepped in as editor and was tireless
in making the process clear and serving as a constant support for the book. We are
grateful for her insight and hard work. Lucy Berbeo, Morgan Shannon, Jennifer Cline,
Alexander Helmintoller, Emma Newsom, and Nathan Davidson all provided much-
needed support in the technical aspects of the text, which are too numerous to mention.
We also thank the people who reviewed earlier drafts of chapters or the whole
book and generously gave us their feedback:
First Edition
Benoit-Antoine Bacon, Bishop’s
University
Mark E. Berg, The Richard Stockton
College of New Jersey
Michael D. Biderman, The University
of Tennessee at Chattanooga
Bryan R. Burnham, The University of
Scranton
Beatrice M. de Oca, California State
University Channel Islands
Robert Dippner, University of
Nevada, Las Vegas
Susan E. Dutch, Westfield State
University
Jim Dykes, University of
Texas at San Antonio
John H. Flowers, University of
Nebraska–Lincoln
Gina A. Glanc, Texas A&M
University Corpus Christi
Paula Goolkasian,
UNC Charlotte
Billy R. Hammond Jr., The
University of Georgia
Cheryl-Ann Hardy, Columbia
College
C. Eric Jones, Regent University
Eric Laws, Longwood
University
Fabio Leite, The Ohio State
University at Lima
Poornima Madhavan,
Old Dominion University
Sara J. Margolin, The College at
Brockport, State University of
New York
Daniel S. McConnell, University of
Central Florida
John Philbeck, The
George Washington University
Elisabeth J. Ploran, Hofstra
University
Jamie Prowse Turner,
Red Deer College
Jennifer S. Queen, Rollins College
Jason A. Rogers, State University of
New York Institute of Technology

xxviii Sensation and Perception
Tom Sanocki, University of South
Florida
Matthew Schlesinger, Southern
Illinois University Carbondale
Carl W. Scott, University of
St. Thomas
T. C. Sim, Sam Houston State
University
Sherril M. Stone, Northwestern
Oklahoma State University
Albert K. Toh, University of
Arkansas–Pine Bluff
Erik C. Tracy, University of North
Carolina at Pembroke
Emily A. Wickelgren, California State
University, Sacramento
Robert J. Woll, Siena College
Takashi Yamauchi,
Texas A&M University
Lonnie Yandell,
Belmont University
Diana L. Young, Georgia College &
State University
Second Edition
Karla M. Batres, Caldwell University
Tifani Fletcher, West Liberty
University
Alexis Grosofsky, Beloit College
Robert J. Hines, University of
Arkansas, Little Rock
Sara J. Margolin, The College at
Brockport, State University of
New York
Nicha K. Otero, University of
Arkansas, Fort Smith
Dennis Rodriguez, Indiana
University South Bend
T. C. Sim, Sam Houston University
Jennifer L. Thomson, Messiah
College
Sue Wilkinson, Cardiff
Metropolitan University
We must also thank many colleagues and students who read drafts of chapters or
who gave us advice on the important questions for a particular topic. These people
include, but are not limited to, Chaz Firestone, Matthew Schulkind, Pat Delucia,
Zehra Peynircioglu, Scott Pearl, Randy Blake, Adrien Chopin, Thomas Sanocki,
Brian Pasley, and Todd Kahan. And thanks to Julia Overton, who sent us information
on chili pepper eating contests.
We thank Hanover College for hosting the ISLE website for the first edition.
Finally, we thank the various individuals who helped out with the art program for
this textbook. Photographs, musical contributions, and likenesses were freely con-
tributed by Margaret Krantz, Leslie Frazier, Sarina Schwartz, Erin Hunter, Jonathan
Altman, Scott Pearl, Robert Kirkpatrick, and Todd Kahan.
Bennett L. Schwartz & John H. Krantz
December 2017

ABOUT THE AUTHORS
Bennett L. Schwartz received his PhD in 1993 from Dartmouth College in New
Hampshire. Since then he has been at Florida International University (FIU) in Miami,
Florida, where he is currently professor of psychology. He is author or editor of 10
published books as well as over 70 journal articles and chapters. His textbook Memory:
Foundations and Applications, 3rd Edition (SAGE), was published in 2017. He has won
several teaching awards at FIU and currently teaches courses in memory, cognition, and
sensation and perception. His main research area is metacognition and memory, but he
has also conducted research in diverse areas that range from visual perception to evolu-
tionary psychology, to the language of thought, and to memory in nonhuman primates.
Schwartz currently serves as the editor-in-chief of New Ideas in Psychology.
John H. Krantz received his psychology PhD from the University of Florida. After grad-
uate school, he worked in industry at Honeywell on visual factors related to cockpit
displays. In 1990, he returned to academia, taking a position at Hanover College. John
has done extensive research in vision, human factors, computers in psychology, and the
use of the Web as a medium for psychological research. He has been program chair and
president of the Society for Computers in Psychology and editor of the journal Behavior
Research Methods. John was the first to develop Web experiments in psychological sci-
ence and led the way on techniques for sending multimedia via the Web. He has served
as a faculty associate for The Psychology Place, developing interactive learning activi-
ties, and created psychology’s first global website for the Association for Psychological
Science (APS). In addition, he is an author for both the Cognitive Toolkit and PsychSim 6.
John is well known for his widely used online psychological experiments related to
sensation, perception, and cognition. His current research is focused on using the Web
for psychological research and modeling the visual system.

1What Is Perception?
Barry Downard/Alamy Stock Photo
LEARNING OBJECTIVES
1.1 Discuss why understanding sensation and perception is important.
1.2 Assess why there are actually more than five senses.
1.3 Describe how transduction transforms a physical signal into a neural signal.
1.4 Illustrate the history of the study of sensation and perception.
1.5 Explain the theory of cognitive penetration, and apply
sensation and perception research to collisions.
INTRODUCTION
Try to imagine the last concert you attended. It does not matter what kind of music you
like—rap, country, even the opera. Most likely, your experience at the concert was thrill-
ing—an experience initiated by the pleasure that auditory perception can bring. The
sounds of the singer’s voice and the guitarist’s riffs stir your emotions. The music pulses
into your ears. Despite your distance from the stage, and the assortment of voices and
instruments, you correctly sort out the lead singer’s voice from
the guitar, the bass, the keyboard, and the drums. Moreover,
even though there is a disconnect between the source of the
sound (the loudspeakers) and the location of the musicians (the
stage), you correctly attribute the sound to the stage. But in
almost all cases, in all varieties of music, a concert is more than
just sound. The flashy clothes and the elaborate video screens
fill your eyes with color and movement. This mixture of sight
and sound is not unique to pop concerts. The same is true for
opera (Figure 1.1). Because opera incorporates drama as well
as music, costumes and visual appearance are as much a part
of the art form as is the music itself. Our enjoyment of a live
concert comes from this coordinated combination of sensory
stimulation.
We tend to take our amazing sensory abilities for granted,
even when we are specifically engaging in activities such as
concerts or visits to art museums, in which we are deliberately
ISLE EXERCISES
1.1 Sequence of Sensory
Events
1.2 Illustration of Action
1.3 Bach’s Violin Partita
No. 2 in D minor
1.4 Column Taper Illusion
1.5 Aristotle’s Illusion
1.6 Motion Aftereffect
1.7 Fechner’s Colors and
Benham’s Top
1.8 Kanizsa Triangle
1.9 Optic Flow
1.10 Kuffler and Single-
Cell Recording
1.11 Size–Arrival Effect
FIGURE 1.1 Chinese Opera
In Chinese opera, elaborate and colorful costumes make
the performance as much a visual experience as it is a
musical one.
©
iStockphoto.com
/xiao-m
ing

2 Sensation and Perception
challenging our sensory systems. In general, we seldom think
about the manner in which we see, hear, touch, taste, and smell
the world, and how we become aware of our own internal
states as well. Perceptual abilities work fast, and they provide
us, during every waking moment, with an ongoing update of
the state of the world around us. We may attend to only cer-
tain aspects of the world at any given moment, but a wealth of
potential information exists around us. Consider the simple act
of reading this paragraph. Unless you are blind and using either
touch or audition to read these words, you are looking at a pat-
tern of black squiggles on a white background. Your attentional
focus is on decoding these squiggles from an intricate pattern
of visual stimulation into meaningful concepts. However, even
if you are in a quiet room, you are surrounded by myriad other
sources of sensory stimulation. Just a slight break of concentra-
tion, and you may become aware of the slight hum of the heat-
ing system, voices from upstairs chatting about campus events,
the pulse of the refrigerator, the Escalade with the deafening
bass line driving by outside, and maybe the plaintive cooing of
a mourning dove in a tree outside the window. Look up from
your text and you can see the flowers your mother sent you, the
leftover dinner that needs to be cleaned up, or your relaxed cat
sleeping by your feet. An inadvertent sniff, smelling the cookies baking in the kitchen,
breaks your concentration (Figure 1.2). Feel free to put “Schwartz and Krantz” down
for a moment and have one of those cookies before you continue reading. And don’t
worry about the vase.
The fact of the matter is that each of these sensory processes, from reading words
in a book to smelling cookies baking in an oven, is incredibly complicated; each one
is completed by exquisitely fine-tuned processes in our brains. By and large, however,
we become conscious only of the finished product—the words on the screen, the
music of the Black Eyed Peas gradually getting louder, then softer, as the car with the
loud stereo passes by outside, the brush of your cat’s soft fur against your skin, and
the taste of the warm chocolate mixing with the sugary cookie dough. The incredible
computing power of the human brain is focused on processing each of these sensory
inputs and allowing your conscious mind to pick and choose among them. Indeed,
sensory systems have evolved to accomplish these processes quickly and efficiently.
This textbook is an introduction to the science of sensory processes. We will
address the issue of sensation and perception from the perspectives of both psychol-
ogy and neuroscience. We will discuss both the neural underpinnings of sensation
and perception and the psychological processes based on those neural processes.
The study of sensation and perception began at the very beginning of scientific psy-
chology, and science has made progress on how sensory processes work. Granted,
there are many questions still unanswered about how sensory experience occurs,
but there is a broad consensus on the general pattern, from how information is
perceived by sensory neurons to the areas of the brain responsible for each sen-
sory system to the cognitive processes whereby we recognize common objects. This
textbook is written to help you understand the complex processes that go into sen-
sory perception. We hold that the study of sensation and perception is a fascinating
area that sheds light on the basic nature of what it is to be human. We hope you find
this journey interesting.
However, at the outset, we find it necessary to inform the reader: Understanding
sensory processes is not easy. No matter how clearly we describe each topic in
©
iS
to
ck
ph
ot
o.
co
m
/S
Co
ut
ts
FIGURE 1.2 Chocolate Chip Cookies
Just seeing chocolate chip cookies may make you hungry.
However, most people cannot make a taste image. Try to
imagine what chocolate chip cookies taste like (not look
like) with nothing actual to taste. Most people cannot
do it. However, when we smell the cookies, we get a
sensation very distinct from what they look like.

3 Chapter 1: What Is Perception?
this book or how well your professor explains the material, understanding this
material takes hard work and patience. The human mind and brain are complex;
therefore, to understand how they work and create our ability to sense and perceive
is also a complex task. Thus, read your text slowly and repeatedly, consult the
online resources, experience the illusions and demonstrations provided on ISLE,
answer the review questions, test yourself repeatedly, and give yourself the time to
learn about sensation and perception. Moreover, the terminology is often complex
and certainly requires ample memorizing. Make sure to learn the key terms, because
they are important in understanding the concepts (and presumably doing well
in your class). Do not get frustrated—it is likely that other students need just as
much time to learn the terms and help understanding sensation and perception as
you do. We think it is worth the time—we hope you will find the incredible designs
and abilities nature has evolved in our sensory systems to be inspiring, interesting,
and important.
WHY IS THIS PSYCHOLOGY?
1.1 Discuss why understanding sensation and perception is important.
You may wonder why sensation and perception are part of the psychology curriculum and
not, say, the biology curriculum. This is a fair question given that much of what you will
read in this book concerns anatomy and physiology, with a little bit of cognitive science
thrown in. The answer, however, is simple: The goal is to understand perceptual experi-
ence, that is, how our brains make sense of the sensory world around us. Understanding
how our minds, through our brains, interpret the world around us is an inherently psycho-
logical goal. We will see, as well, that psychological processes, such as attention, intention,
emotion, and biases, influence the ways in which we perceive the world. Consider the
sensory experience of pain. Every marathon runner, whether an amateur or an Olympic
medalist, crosses the finish line in intense bodily discomfort (Figure 1.3). Body tempera-
ture is at fever levels, muscles are filled with lactic acid,
sunburn covers the skin, dehydration is present, and sores
are opening on the feet. But despite the pain and discom-
fort, the marathon runner is likely ecstatic—he or she
has completed an incredibly challenging goal. The pain
is interpreted in the context of completing a long-sought
objective—to demonstrate that the marathon can be run.
If the same physical conditions were duplicated in a prison
setting, it could be interpreted as torture.
Consider, too, how our personal biases can influence
perception. Think about two friends from Los Angeles
watching a basketball game between the Los Angeles
Clippers and the Los Angeles Lakers. Each person is
watching the same game from the same vantage point in
the stands. That is, both friends are receiving the same
perceptual input. However, each person perceives dif-
ferent things. The Clippers fan sees Lakers star Lonzo
Ball “charging,” that is, committing an offensive foul.
The Lakers fan sees the same play as a blocking foul
on Clippers star DeAndre Jordan, maybe even a flagrant
©
iStockphoto.com
/bytepark
FIGURE 1.3 Runner Overcoming Pain
Marathon runners must overcome intense pain by the end of
the race. Our bodies have a number of ways of easing the pain.
The marathon runner may value the sense of accomplishment of
finishing the race. But pain is still an intense sensory experience.

4 Sensation and Perception
foul. The same sensory input results in two different perceptual outputs. Research has
shown that even when participants are instructed to be totally objective, they cannot
overcome the kinds of biases that are created by being a fan of one team or another
(e.g., Hastorf & Cantril, 1954). That is, even when trying to be totally objective, Lakers
fans cannot see the game from a Clippers perspective, and Clippers fans cannot see the
game from a Lakers perspective. For this reason, referees who grew up as Lakers fans
are seldom assigned the job of refereeing games involving Los Angeles teams. The hope
is that a referee from Seattle can see a game in an unbiased way (or at least a less biased
way) compared with a referee from Los Angeles. However, because of the biases of the
Clippers fans and the Lakers fans, they will never see what the referee sees. Thus, the
fairest of referees may often be scolded by fans of both teams as being biased toward
one or the other team—in the very same game.
Knowledge can influence perception in more subtle ways than occur among
biased sports fans. Consider seeing a small dot on the horizon of the ocean. Because
you know that it is a cruise ship out there, you tend to see it as such. That the object
could be an alien spaceship is considered far less likely, so you do not see it as such.
Indeed, we use our knowledge of objects to help us perceive them even in relatively
simple and static scenes.
TEST YOUR KNOWLEDGE
1. Why are sensation and perception studied by psychologists?
2. How do personal biases influence perception? Do such biases influence what
we see and hear, or do they bias what decisions we make about what we see
and hear?
THE MYTH OF FIVE SENSES
1.2 Assess why there are actually more than five senses.
Most of us were taught in elementary school about the five
senses—vision, hearing, touch, smell, and taste. This standard
taxonomy is not a simplified classification system, though it
goes all the way back to Aristotle. It is wrong! We certainly
have more than five senses. For example, in addition to these
sensory systems, we have a vestibular system to help keep our
balance and a proprioception system to allow us to monitor the
position of our bodies. Our sense of touch is composed of mul-
tiple physiological systems designed to sense different features
of the environment. Heat, cold, pain, itchiness, and soft touch
are all implemented by separable systems. Indeed, the receptors
for the “itch” experience are a unique kind of receptor different
from those that sense touch and those that sense pain. Thus,
depending on how the different touch systems are counted, it is
more realistic to say that human beings have anywhere from 7
to 12 different sensory systems (Table 1.1). Indeed, some have
even argued that hunger and thirst should be counted as senses.
We leave them out, as they deal strictly with internal states that
are not directly linked to perception of the external world.
TABLE 1.1 Senses of the Human Body
This table shows that there are more sensory systems than the
traditional five.
Function Organ
External
or Internal
Stimuli
Vision Eyes External
Hearing Ears External
Smell Nose External
Taste Tongue External
Light touch Skin External
Pressure Skin External
Cold Skin External
Heat Skin External
Pain Skin/viscera External/internal
Itch Skin External
Vestibular Inner ear External
Proprioception Muscles Internal

5 Chapter 1: What Is Perception?
The standard five-sense approach also fails if you think of the sensory systems as
hierarchically organized. For example, our perception of the flavors of foods is a com-
plex interaction of smell and taste, and it is something we feel is quite different from
smell itself. Flavor may also take into account other sensory modalities, such as vision.
For example, if you have ever tried to eat green eggs and ham, you know what we are
saying. However, because smell and taste are so bound together, it makes sense to group
these two senses in the same way that soft touch, pain, and temperature are grouped
together as the sense of touch or as a somatosensory system. Nonetheless, old habits are
hard to break, and this book is roughly organized according to the traditional five-sense
taxonomy. This is merely a bow to convention, not an endorsement of any scientific truth
behind it. In Chapter 14 we cover the skin senses (touch, pain, and temperature). You
will see that the various senses that make up this group are different both perceptually
and anatomically. Indeed, even the sensation of “hotness” and the sensation of “cold-
ness” are brought about by different receptor systems, not a differential response from a
“temperature” receptor. The classic “burn” sensation of feeling liquid nitrogen on one’s
skin is an example of this (removing of warts and moles with freezing liquid nitrogen is
a common dermatological procedure).
It makes scientific sense to discuss the visual system and the auditory system as
discrete sensory systems. And given the centrality of these two systems to human
sensory processing, much of this book is devoted to them. Indeed, all but the last two
chapters are about vision and audition. Increasingly, however, how sensory systems
interact is becoming an important topic of study known as multisensory processing,
studies of which address how one sense can affect perception in another (Alvarado,
Vaughan, Stanford, & Stein, 2007).
TEST YOUR KNOWLEDGE
1. What is meant by the myth of five senses?
2. What is the difference between internal and external senses? Can you give an
example of each?
THE BASICS OF PERCEPTION
1.3 Describe how transduction transforms a physical signal into a neural signal.
Sensation refers to the registering of a physical stimulus on our sensory receptors.
That is, sensation is the earliest stage of a process that starts off in the eyes, ears, or
skin and ends in the higher centers of the brain. Sensation changes physical stimuli,
such as light, sound waves, and mechanical vibrations, into information in our nervous
systems. Perception, by contrast, refers to the later aspects of the perceptual process.
To be specific, perception involves turning the sensory input into meaningful conscious
experience. In this sense, perception means the translation of that neural signal into
usable information.
Sensation and perception are usually thought of as distinct processes, one refer-
ring to the basic process of converting external information into a neural signal
and the other concerned with interpreting what that signal means. Other research-
ers think that the dichotomy between sensation and perception is a false one (e.g.,
Gibson, 1979). We will begin with the standard model, in which sensation occurs in
the sensory organs. For audition, sensation occurs in the ear, in particular, the hair
1.3
Sensation: the registration
of physical stimuli on sensory
receptors
Stimulus: an element of the
world around us that impinges
on our sensory systems
Perception: the process of
creating conscious perceptual
experience from sensory input

6 Sensation and Perception
cells of the cochlea. For our sense of touch, sensation occurs along the surface of the
skin. Perception occurs after cognitive processing begins, typically in the cerebral
cortex of the brain. Sensation is about stimuli; perception is about interpretation.
To perceive the world, we need both. We cannot perceive the Russian doll on the
coffee table as a Russian doll on the coffee table without both ends of the process
(Figure 1.4). The image of the doll must fall on the retina in order for us to perceive,
but equally important is the parsing of the perceptual environment, that is, knowing
where the doll ends and the table begins. Perception also involves knowing what you
are looking at.
So let’s take a look at the perceptual process, that is, the sequence of mental
operations that bring us from the initial sensory input to our understanding of our
conscious experience (ISLE 1.1).
Stimuli reflect light, produce sounds and vibrations, have surface texture, and
produce volatile chemicals (which we can smell). The job of our perceptual processes
is to determine what is out there in the world around us. We want our perceptual
processes to produce a veridical (truthful) representation of what surrounds us and
allow us to focus on those stimuli that are most important to us. Veridicality is
important because we want our sensory systems to be guiding us in adaptive ways.
For example, if we perceive an object as being farther away than it actually is, we
may bump into it and injure ourselves.
Because there are many potential stimuli in our world, we must be able to focus on
potential stimuli that are important or interesting. Such stimuli are called the attended
stimuli. When most of us listen to a song, our attention is drawn to the voice of the
singer or the melody of an instrumental tune. We can pick out different parts of the
music, say, the drums or the bass, by switching the attended stimulus from the lead
singer to the bass line, but usually our attention is focused on the melody. With vision,
we may be looking at a beautiful beach, but our attended stimulus may be the beer
bottle a previous beachgoer left behind.
Through our senses, we are presented with an incredibly rich and varied experi-
ence of the world, including the aroma of roasting coffee, the texture of fine silk, the
taste of dark chocolate, the sound of our favorite musician, and the sight of a glorious
sunset. Not all sensory experiences are pleasant and lovely,
of course. We have all smelled rotting garbage, felt a painful
pinprick, placed our fingers on a hot stovetop, tasted foods
we detest, heard fingernails screeching on the blackboard, and
seen images in movies that have made us close our eyes. The
senses unflinchingly bring to us an immense range of experi-
ences from the world around us, positive and negative alike.
Yet the next question we can ask is how do stimuli in
the outside world become perceptual experiences? How do
our sensory systems, such as our eyes, ears, and skin, trans-
late light, sound, and surface textures into perceptions? For
example, how does our nose turn the volatile chemicals of
coffee into the tantalizing aroma of coffee (Figure 1.5)? We
need to explore how physical stimuli are converted into a
sensory representation. So we now introduce the concepts of
transduction and neural responses.
For each of our sensory systems, we have specialized
neural cells called receptors, which transduce (transform)
a physical stimulus into an electrochemical signal, called a
neural response, which then can be sent to the brain. For
FIGURE 1.4
Seeing: A Complex Process
To understand this photograph,
you must sense the colors and
images, but cognitive processes
aid in understanding what it
is you are looking at. Without
some cultural knowledge, the
sensation makes little sense.
However, to our conscious
selves, the process of perceiving
the Russian dolls is seamless.
©
iS
to
ck
ph
ot
o.
co
m
/S
yl
vi
eB
ou
ch
ar
d
ISLE 1.1
Sequence of Sensory Events
FIGURE 1.5 The Smell of Coffee
The smell of coffee comes from molecules in the air that
rise from the coffee. Special cells in our noses must convert
the presence of those molecules into neural signals, which
we interpret as the wonderful aroma of coffee.
©
S
hu
tt
er
st
oc
k/
Ra
ch
at
a
Te
yp
ar
si
t

7 Chapter 1: What Is Perception?
FIGURE 1.6 Converting Energy Into a Neural Signal
Perception is the process of converting physical stimuli, such as light and sound energy, into neural signals within our sensory organs.
(a) Light is reflected off the petals of the flowers and into the eyes. The eyes then transduce this light into a neural signal to be sent to
the brain. (b) Sound is produced by the bird and reaches our ears. Special hair cells in the cochlea of the ear transduce the sound into
a neural signal to be sent to the brain.
vision, these specialized cells are called rods and cones, and they are located on the
retina of the eye. For hearing, these specialized cells are called the hair cells, and
they are found in the cochlea in the inner ear. For our sense of taste, we have cells
in our taste buds, which create neural signals when they come into contact with
certain chemicals in our food. Rods and cones transduce the physical energy of light
into an electrochemical signal, which is then transmitted to the brain via the optic
nerve. Hair cells convert the vibrating of the cochlear membrane (which vibrates in
response to physical sound) into a neural response, which is then transmitted to the
brain via the auditory (or cochlear) nerve. Taste bud cells convert the presence of a
particular chemical (such as sugars) into a neural response transmitted to the brain
by gustatory nerves (Figure 1.6).
Once a neural signal is transduced by the receptors, it is transmitted to the brain
for processing. Though the neural signal contains much information, it is necessary
for the brain to process that information in order to extract relevant information,
such as color for vision and pitch for sound. It is for this reason that we find it useful
to distinguish between sensation and perception. Sensation refers to the process of
transduction, in which receptors convert physical signals into neural responses, and
perception refers to the process of taking that signal and processing it into a usable
image or experience. For example, when we hear orchestral music, the hair cells in
our cochleae convert the sound waves into neural signals, but it is our brains that
convert that neural signal into the experience of the music we hear, the rich sound of
Transduction: the process of
converting a physical stimulus
into an electrochemical signal
Receptors: specialized sensory
neurons that convert physical
stimuli into neural responses
Neural response: the signal
produced by receptor cells that
can then be sent to the brain
(a) (b)

8 Sensation and Perception
the violins contrasting with the sharp tones of the trumpets, and
the underlying low tones of the basses and bassoons. Sensation
enables us to take in the sounds, but perception allows us to
appreciate the music.
Action
The goal of sensation and perception is to guide us through our
environment. We use visual information to avoid obstacles while
walking around campus or driving on the highway. When it is
completely dark, we avoid obstacles by touch. We feel our way by
touching the obstacles in front of us. In this way, most of us can
negotiate our own homes at night with the lights off just using
touch. We use auditory information to determine what people are
saying and whether the phone is ringing. We use olfactory (smell)
information to determine if we want to eat something or avoid
any contact with it. We use information about temperature on our
skin to determine if we want to wear a sweater or a T-shirt. Thus,
perceiving what is around us guides us to action. We can define action as any motor
activity. Thus, action includes moving one’s eyes along the page of a book as well as a
baseball player’s swinging his bat at an incoming fastball (Figure 1.7). It includes turning
your head when you hear the voice of a friend or a concert pianist’s fingers darting across
her piano keyboard. Thus, any movement we make can qualify as action. If that move-
ment is directed by something we perceive in the environment, then we can see that it is
perception-guided action. In sum, one of the chief goals, if not the
goal itself, is to guide functional action. (There are some interesting
illustrations of this on ISLE 1.2.)
The Nature of Experience
and Phenomenology
Think about the experience of listening to very beautiful music. For
example, think of a violin virtuoso playing Bach’s Violin Partita
No. 2 in D minor. If you are not familiar with this piece of music,
you can listen to an excerpt of it (ISLE 1.3). If you do not like this
style of music, imagine a piece of music that you think is very sad
but also very beautiful. In the Bach partita, our ears are respond-
ing to differences in frequency, tempo, rhythm, and loudness as we
hear the sound of the lone violin. Our toes may move in time to
the music, and our eyes may cloud with tears at the poignancy of
the music. These movements may be considered action. But
why does the piece affect us emotionally? Why do we find music beautiful in the first
place? How is it that we feel Bach’s grief at losing his wife through his music nearly 300
years after the piece was written? If you are a music theorist, this speaks to the power
of Bach’s music. But for our purposes, it introduces us to the issue of phenomenology.
Phenomenology is our subjective experience of perception. Phenomenology refers to our
internal experience of the world around us. Phenomenology is the beauty of the lone vio-
lin, the aromatic smell of coffee in the morning, and the wonder of seeing the colors of
the sunset in the west of an evening sky (Figure 1.8). By the same token, phenomenology
can refer to the annoying cacophony of the neighbor’s lawnmower, the stink of an airplane
ISLE 1.2
Illustration of Action
FIGURE 1.7 Perception and Action
The baseball player’s swing is guided by the perception
of the approaching baseball. Opposing pitchers may
throw curveballs, which change direction at the last
moment, to confuse batters’ perceptions.
©
iS
to
ck
ph
ot
o.
co
m
/R
BF
rie
d
FIGURE 1.8 Phenomenology
When we see a beautiful sunset, we notice the colors
and the landscape. The experience of all this is
considered its phenomenology. Issues of phenomenology
interest psychologists and philosophers alike.
©
iS
to
ck
ph
ot
o.
co
m
/t
om
of
bl
ue
sp
rin
gs
Action: any motor activity
Phenomenology: our
subjective experience of
perception

9 Chapter 1: What Is Perception?
bathroom, and an up-close look at a stranger’s nose hairs. Regardless of the effects they
induce in us, perception induces these internal mental experiences in each of us.
Phenomenology distinguishes us from computer-driven robots. Robots have
microphones to capture sound and internal computer processing units to decode
the messages into something that can drive the robots to follow certain courses of
action. But computer devices do not have internal experiences (as best we know).
Phenomenology appears to be a unique creation of the living brain. Philosophers
have argued about what purpose it serves, whether people share common phenom-
enology, how widespread it is among nonhuman animals, and if it is possible in
nonliving systems. The discussion of these questions is largely outside the scope of
this book, but there is a lively debate in philosophy concerning how we would know
if animals or artificially intelligent devices experience phenomenology when they
process the external world.
An interesting issue in phenomenology is that it is a private experience. Each of us
has sensory experiences, and we agree on common terminology for them, but because
phenomenology is private, philosophers often wonder if this means that we share phe-
nomenology. For example, scientists know what frequencies of light elicit an experi-
ence of blue, and people all across the world agree on what constitutes blue (regardless
of language). We may also share common cultural referents to blue, which we all agree
on (blue jeans, blue moods, blue states, etc.). However, is our internal experience of
blue the same? Do you and I have the same phenomenology of blue? This conundrum
is often referred to as the inverted-rainbow question. We all agree on the relation of
frequency to color name when we see a rainbow, but what if our internal experiences
were different? By and large, this question takes us out of the domain of scientific psy-
chology, as it cannot be answered empirically. But philosophers take these questions
quite seriously. At sagepub.com, you can find some approachable readings on the phi-
losophy of phenomenology.
TEST YOUR KNOWLEDGE
1. What is transduction? How might it differ from one sensory system to the next?
2. What is phenomenology? Why do you think it is so difficult to study by
experimental psychologists?
THE HISTORY OF SENSATION
AND PERCEPTION
1.4 Illustrate the history of the study of sensation and perception.
To understand any field of study, it is helpful to understand its history. Often the ques-
tions asked in the past and the way they were posed influence modern research. Therefore,
knowing the history can often help us understand why issues are framed the way they are
and why our knowledge has both the strengths and the weaknesses that it does. It can
also help us understand our own assumptions about the nature of the world by seeing
ideas from different epochs, when ideas and technology were different. Sometimes a field
does not even make sense until its history is considered. In this section, a short history of
the field of the study of sensation and perception is covered to give some context to the
material in this chapter and throughout the book.
ISLE 1.3
Bach’s Violin Partita
No. 2 in D minor

10 Sensation and Perception
The Beginnings
Thinking about our senses probably predates any written record. We
would like to think that even in prehistoric times, people marveled at
the beauty of a rainbow or a sunset. The formal study of sensation
and perception goes nearly as far back as written records exist. The
Ramesseum medical papyri date back to approximately 1800 BCE.
That’s nearly 4,000 years ago! The unknown authors describe disrup-
tions in visual perception and their connection to diseases of the eye
(Weitzmann, 1970). Interestingly, the papyri recommend hemp as a
treatment for eye disorders, much as doctors now prescribe medical
marijuana for glaucoma. In ancient Greece (some 1,500 years after the Ramesseum
medical papyri), Greek architects were aware of how perception could be distorted by
visual illusions (Coren & Girgus, 1978). Indeed, the Parthenon was built to appear as if
it had straight columns, though in order to do so, the architects had to design columns
that were not perfectly vertical. Elsewhere, Greek architects built perfectly straight col-
umns, but these appear bent (ISLE 1.4).
Aristotle (384–322 BCE) conducted conceptual work and observations in the field
of sensation and perception. He clearly distinguished between sensory and motor
functions; he described the sensory organs and their functions; he even gave us our
prototypical list of five senses: sight, hearing, smell, taste, and touch (Murphy &
Kovach, 1972). In addition to these basic ideas, Aristotle was a keen
observer and is the first to have recorded two very interesting sensory
phenomena, which we describe here, as they are relevant to topics we
discuss in later chapters.
The first is an illusion of touch that still bears his name, the Aristotle
illusion (Benedetti, 1985). In this illusion, a single touch between the tips
of two crossed fingers, say with a pen, will be experienced as if there were
two touches, as if there were two pens and not a single one (ISLE 1.5).
You can experience the illusion very simply by crossing your index and
middle finger (Figure 1.9). Then pick up a pencil and touch the place just
where the two fingers meet. You probably experience two touches (as
there is one touch to each finger) that feel like two pencils. The Aristotle
illusion is relevant to material discussed in Chapter 14.
The second illusion discussed by Aristotle is known as the motion
aftereffect, also known as the waterfall illusion (Verstraten, 1996; Wade,
1996). A motion aftereffect is a sensory experience that occurs after
prolonged experience of visual motion in one particular direction (ISLE
1.6). So if we think of a waterfall, the water is all moving in the same
direction: down. That is the only way real waterfalls actually occur.
After staring at a waterfall for some time, if you move your eyes to a
solid stationary object, you may get the sense that the stationary object
is moving upward, that is, in the direction opposite the one in which the
waterfall was moving. Another way to observe this illusion at home is to
watch the credits at the end of a movie without taking your eyes off the
television screen. After watching the credits for 2 minutes, have a friend
or family member hit the pause button. You will see an illusion of the
movie credits’ moving up the screen, even though you know the video is
now stopped. Aristotle observed this by looking at nonmoving surfaces
after watching the downward motion of waterfalls (Figure 1.10).
Many philosophers and scientists prior to the 19th century devel-
oped ideas about perception, some of which still influence our thinking
ISLE 1.5
Aristotle’s Illusion
ISLE 1.6
Motion Aftereffect
FIGURE 1.9 The Aristotle Illusion
In the illustration, we see two crossed fingers and
a pencil touching in the middle. In this illusion, we
feel as if we have been touched by two pencils
rather than one.
FIGURE 1.10 Waterfall Illusion
After watching the constant downward motion
of a waterfall, the motion detectors in our brains
adapt or tire in response to the downward
motion. When we look away from the waterfall,
upward motion detectors become active, and we
experience the illusion that whatever stationary
object we are looking at is moving upward.
©
iS
to
ck
ph
ot
o.
co
m
/M
ax
le
vo
yo
u
ISLE 1.4
Column Taper Illusion

11 Chapter 1: What Is Perception?
today. For example, the astronomer Robert Hooke (1635–1703) developed the first
acuity test for vision (Grüsser, 1993). However, it is with the 19th century that we
see the beginnings of a real science of sensation and perception. The contributions
came from people in many different disciplines. Physicist Thomas Young (1773–1829)
argued that light is a wave and that color is detected by three kinds of nerve fibers,
an idea elaborated on later by Hermann von Helmholtz. Biologist Johannes Mueller
(1801–1858) developed the doctrine of the specific nerve energies. The doctrine of
specific nerve energies argues that it is the specific neurons activated that determine
the particular type of experience. That is, activation of the optic nerve leads to visual
experiences, whereas activation of the auditory nerve leads to auditory experiences.
This seemingly obvious conception was controversial at the time because one of the
confusing findings of that time period was how similar the electrical activity was in all
of the neurons, regardless of the sensory modality. This led Mueller and others to won-
der how the brain could distinguish between seeing an apple and hearing a song, for
example. Those neurons involved in seeing will cause the impression of seeing regard-
less of how they are stimulated. So if a sound stimulates the neurons involved in vision,
you will still have a visual experience. Some early evidence for this view is the visual
experience we have when we manually stimulate the retina. Try pushing at the side of
your eye, and you will see that pushing on the eye can cause shifts of color and shading.
Helmholtz Versus Hering
One of Mueller’s students was Hermann von Helmholtz (1821–1894) (Figure 1.11).
Helmholtz was born in Potsdam in what was then the Kingdom of Prussia (now part
of Germany). He was a professor for many years at the University of Berlin. Helmholtz
was trained in medicine and physiology but became famous for his work in both sen-
sory perception and physics. His work in physics includes the formulation of the law
of the conservation of energy, an important landmark in physics. He was also a pioneer
in thermodynamics and electrodynamics. In biology, Helmholtz was the first person to
determine the speed of the neural impulse (or action potential). He also contributed
greatly to the beginnings of sensory physiology and the scientific study of perception.
Indeed, Helmholtz wrote a three-volume study of physiological optics and also wrote a
book on music perception. In particular, he elaborated on the work of Thomas Young
and developed the concept of the trichromatic theory of color vision. For this reason,
the trichromatic theory of color vision is often called the Young–Helmholtz theory of
color perception. Helmholtz thought of our color vision as being based on the percep-
tion of three primary colors, red, green, and blue, even though he was not certain that
there were three cone types in the retina (this was not known for certain until much
later). By and large, his color theory is still relevant today.
Helmholtz developed a general theory of how our senses work, which is mostly
still held today by most researchers within the field of sensation and perception. In his
constructivist approach, Helmholtz argued that the information from the sensory sig-
nal itself is inadequate to explain the richness of our experiences. That is, the sensory
signal needs to be interpreted by active cognitive processes. For example, recognizing
the face or voice of a loved one is more than basic sensations. Your auditory system
must integrate the sound of the voice with your knowledge of your sister and the
knowledge of what she usually talks to you about in order to fully perceive the speech
directed at you. Thus, we must incorporate information from our existing knowledge
to completely perceive the world around us. According to Helmholtz, because our
senses do not produce sufficient information about the world, we must use a form
of reason, unconsciously, to make an educated guess about what we actually per-
ceive. Helmholtz called this an unconscious inference (Turner, 1977). This type of
Aftereffect: a sensory
experience that occurs after
prolonged experience of visual
motion in one particular direction
Doctrine of specific nerve
energies: the argument that it
is the specific neurons activated
that determine the particular
type of experience
Constructivist approach:
the idea that perceptions are
constructed using information
from our senses and cognitive
processes
Unconscious inference:
perception is not adequately
determined by sensory
information, so an inference
or educated guess is part of
the process; this inference
is not the result of active
problem solving but rather of a
nonconscious cognitive process
FIGURE 1.11
Hermann von Helmholtz
Hermann Ludwig Ferdinand von
Helmholtz (1821–1894) was a
German physician, physicist,
and sensory physiologist. He
is credited with developing
a theory of color vision and
promoting the constructivist view
of sensory perception.
©
G
L A
rchive/A
lam
y

12 Sensation and Perception
theory is useful for explaining the occurrences of auditory and visual illusions, such
as the waterfall illusion. In this way, Helmholtz’s work foreshadowed the cognitive
approaches known as information processing and the computational approach.
Helmholtz’s colleague and rival was Ewald Hering (1834–1918) (Figure 1.12).
Hering was born in Saxony, in what is now Germany, but he was a professor in
Prague in what is now the Czech Republic and Leipzig in Germany. Hering dis-
agreed with Helmholtz both on the specifics of color vision and on a general model
of how sensory processes worked. With respect to color vision, Hering did not see
color vision as being based on three primary colors but as being based on color
opponency (Turner, 1993). He saw two major pairs of color opponents, green–red
and blue–yellow. Any receptor excited by green turns off red, and any receptor
excited by red turns off green. Modern research suggests that both Helmholtz and
Hering were correct to some extent. Trichromacy seems to best explain the work-
ings of the retina, whereas opponency can account for how areas of the visual brain
(occipital cortex) treat color. That they were both right would have displeased both
of these proud scientists, who deeply respected each other but were also fiercely
competitive with one another. This controversy is discussed at length in the chapter
on color perception (Chapter 6). Look ahead to the demonstration on this issue in
that chapter.
Hering also disagreed with Helmholtz’s theory of unconscious inference. Hering
viewed environmental inputs and our sensory apparatus as sufficient for us to grasp
the structure of the perceived world, without the need for internal unconscious
inferences. That is, stimuli contain adequate information for viewers to perceive
the world. In Hering’s view, the perceptual processes in the brain do not need to
make sense of the perceptual world; the brain simply needs to register it. Whereas
Helmholtz’s view is more popular with most experimental psychologists as well as
physiologists, Hering’s view influenced the development of gestalt psychology and,
later, direct perception theory (also known as the Gibsonian view).
Weber, Fechner, and the Birth of Psychophysics
Helmholtz and Hering both approached sensation and perception from the perspective
of physiology. Around the same time as Helmholtz and Hering were looking at the
relation of physiology and perception, other German scientists were doing work with a
more psychological perspective. Ernst Heinrich Weber (1795–1878) discovered Weber’s
law (though it was Gustav Fechner, another German scientist, who named the law after
Weber). Weber’s law states that a just-noticeable difference (JND) between two stimuli
is related to the magnitude or strength of the stimuli. What does this mean? Well, it con-
cerns two stimuli that are very similar. Can we detect the difference between two very
close red colors, or can we detect the difference between 1.44 mg of sugar dissolved in
a cup of water and 1.48 mg of sugar dissolved in a cup of water? Thus, Weber’s law
concerns the perception of difference between two stimuli. For more examples, do we
hear the difference between a 1,000-Hz tone and one of 1,005 Hz? Another example
is whether we see the difference in length between a line 466 mm long and one that is
467 mm long. Weber’s law suggests that we might not be able to detect a 1-mm dif-
ference when we are looking at lines 466 and 467 mm in length, but we may be able
to detect a 1-mm difference when we are comparing a line 2 mm long with one 3 mm
long. Another example of this principle is that we can detect 1 candle when it is lit in
an otherwise dark room. But when 1 candle is lit in a room in which 100 candles are
already burning, we may not notice the light from this candle. The JND is greater for
very loud noises than it is for much more quiet sounds. When a sound is very weak, we
can tell that another sound is louder, even if it is barely louder. When a sound is very
loud, to tell that another sound is even louder, it has to be much louder. Thus, Weber’s
Weber’s law: a just-noticeable
difference between two stimuli
is related to the magnitude or
strength of the stimuli
FIGURE 1.13
Gustav Fechner
Gustav Fechner (1801–1887)
is considered the “father of
psychophysics.” His landmark
work on the relation between
physical stimuli and perception
established sensory psychology as
a unique discipline separate from
physiology. His work inspired the
beginning of scientific psychology.
D
ig
ita
l L
ib
ra
ry

S
m
ith
so
ni
an
In
st
itu
tio
n
Li
br
ar
ie
s
FIGURE 1.12
Ewald Hering
Karl Ewald Kostantin Hering
(1834–1918) was a German
physiologist who introduced
the opponent theory of color
vision. He was also interested
in binocular vision. He
disagreed with Helmholtz on
constructivism. Hering argued
that stimuli themselves had
sufficient information to allow for
direct perception.
U
.S
. N
at
io
na
l L
ib
ra
ry
o
f M
ed
ic
in
e,

H
is
to
ry
o
f M
ed
ic
in
e
D
iv
is
io
n

13 Chapter 1: What Is Perception?
law means that it is harder to distinguish between two samples when those
samples are larger or stronger levels of the stimuli. To repeat, Weber’s law
states that a “just-noticeable difference” (JND) between two stimuli is related
to the magnitude or strength of the stimuli.
Gustav Fechner (1801–1887) is generally considered the founder of
psychophysics, the study of the relation between physical stimuli and
perception (Figure 1.13). Fechner’s (1860/1966) book Elements of
Psychophysics is often considered the beginning of the psychological study
of sensation and perception. Fechner discovered the illusion known as
the Fechner color effect, whereby moving black-and-white figures create
an illusion of color (ISLE 1.7). This illusion is also known as Benham’s
top (Figure 1.14). Fechner also developed Fechner’s law, which states that
sensation is a logarithmic function of physical intensity. This means that
our sensory experience changes at a lower rate than does the physical
intensity. That is, our perception of the intensity of a stimulus increases at
a lower rate than does the actual intensity of the stimulus. For example,
his law is captured in the decibel scale that we use to measure loudness, on
which 20 dB is 100 times louder than 10 dB in terms of the physical stim-
ulus. But in psychological studies, we hear 20 dB as only twice as loud as
10 dB, not 100 times. In vision, the psychological concept of brightness is
a function of the intensity of a light. But to perceive a doubling of bright-
ness, the intensity of the light must increase 10-fold. Fechner’s book and
his view on the relation between physical stimuli and psychological per-
ception influenced many early psychological scientists, including Wilhelm
Wundt, Hermann Ebbinghaus, and William James.
The 20th Century and the Study of
Perception: Cognitive Psychology Approaches
The 20th century brought a burst of interest and research in the study of sensation and
perception. First, sensation and perception research spread from Germany to many
other countries across the world, and a number of perspectives emerged, including
gestalt psychology, direct perception (or the Gibsonian approach), the information-pro-
cessing approach, and the computational approach. These latter two, information
processing and the computational approach, can be considered cognitive psychology
approaches. We briefly describe each approach here.
Gestalt Psychology
Gestalt psychology in general and sensation and perception in particular argued that
we view the world in terms of general patterns and well-organized structures rather than
separable individual elements (Schultz & Schultz, 1992). Consider Figure 1.15. In this
figure, we see the A only when we order the smaller elements together. In another exam-
ple, gestalt psychologists were interested in apparent motion, which can be explained
only by reference to the interaction between parts, not the individual parts themselves.
Gestalt psychologists were also interested in how edges are perceived, an interest that
has continued in both computational and neuroscience approaches to sensation and
perception. Gestalt psychologists considered the visual perception of edges as critical
to determining what objects were. They also identified several situations in which we
see illusory edges on the basis of gestalt principles. One of these is the famous Kanizsa
triangle, depicted in Figure 1.16 (ISLE 1.8). In the Kanizsa triangle, we see illusory
contours, which are suggested by the overall pattern of the figure but are not physically
there. The gestalt psychologists established a number of laws, which they argued were
ISLE 1.7
Fechner Colors and Benham’s Top
Psychophysics: the study of
the relation between physical
stimuli and perception events
Gestalt psychology: a school
of thought claiming that we view
the world in terms of general
patterns and well-organized
structures rather than separable
individual elements
FIGURE 1.14 Benham’s Top
This illusion was discovered by Fechner. Copy
the image shown here. Or look up Benham’s
tops on the Internet and print out a copy of the
above image. Put a pin or needle through the
very center of the image. Then spin the image
as fast as you can. While the image is moving,
you may see colors. Stop the movement and
look at the image again. The colors will be gone.
FIGURE 1.15
Gestalt Psychology
To see the A, one must order the
individual elements into a pattern.
By examining only the individual
elements, we would never see
the A. Gestalt psychologists
considered patterns such as these
the rule rather than the exception.
X X
X
X X
X X
X X
X X X X X X
X X
X X
X X

14 Sensation and Perception
constants in visual perception. These laws were devised to explain how patterns are
seen from individual elements (Figure 1.17).
Gestalt psychology flourished in Europe in the early 20th century. However, during
the same time period, American psychology was under the influence of behaviorism,
which had a very different perspective than did gestalt psychology. Behaviorists were
firm believers in the importance of “nurture” when it came to developmental issues
in psychology, whereas gestalt psychologists were firmly in the “nature” camp; that
is, they believed that these perceptual laws and other principles of human behavior
were genetically wired. Thus, gestalt psychology did not take hold in the United States.
However, when the Nazis came to power in Germany, many gestalt psychologists came
to the United States and Canada, where they influenced the development of the direct
perception view of perception, which we discuss shortly.
ISLE 1.8
Kanizsa Triangle
FIGURE 1.16
Kanizsa Triangle
When most people look at
this figure, they see a bright
white triangle lying on top of a
background consisting of a less
bright triangle and some odd-
shaped “Pac-Man” figures. The
bright white triangle is illusory.
The triangle is suggested by
the pattern of figures, and our
perceptual systems enhance
the perceived brightness of the
figure, but close inspection of the
figure will convince you that there
is no actual change in brightness.
Ka
ni
zs
a
Tr
ia
ng
le
FIGURE 1.17 The Laws of Gestalt Psychology
a. The law of proximity. You will see this
arrangement as a set of columns–not a set
of rows. Items that are near each other are
grouped together. Now notice the typing in
this book. You see rows of letters rather than
columns because a letter is closer to the
letters to the right and left than it is to the
letters above and below.
c. The law of good continuation. You will see
a zigzag line with a curved line running
through it, so that each line continues in the
same direction it was going prior to intersection.
Notice that you do not see the �gures as being
composed of the two elements below.
d. The law of closure. You will see a circle
here, even though it is not perfectly closed. A
complete �gure is simply more tempting than
a curved line! Now close this book and put
your �nger across one edge, focusing on the
shape of the outline of your book. You should
still see your book as complete, but with a
�nger in front of it.
e. The law of common fate. If dots 1, 3, and 5
suddenly move up and dots 2, 4, and 6—at
the same time—suddenly move down, the
dots moving in the same direction will be
perceived as belonging together. The next
time you look at automobile traf�c on a
moderately busy street, notice how clearly
the cars moving in one direction form one group
and the cars moving in the opposite direction
form another group.
Look out the window at the branches of a
tree, and focus on two branches that form a
cross. You clearly perceive two straight lines,
rather than two right angles touching each
other.
b. The law of similarity. You will see this
arrangement as a set of rows rather than
columns. Items that are similar to each other
are grouped together. Now look at the two
words at the end of this sentence that are in
boldface type. Notice how these two words
in heavier print cling together in a group,
whereas the words in regular, lighter print
form their own separate groups.
+ + + + + + +
+ + + + + + +
+ + + + + + +

15 Chapter 1: What Is Perception?
One prominent gestalt psychologist was Wolfgang Köhler (1887–1967). Köhler
was a patriotic, aristocratic German, who nonetheless vehemently opposed the Nazis
(Figure 1.18). In 1933, he was the last person in Germany to publish an article
critical of the Nazis during their regime. Soon afterward, he escaped to the United
States. Once there, he became a successful professor first at Swarthmore College in
Pennsylvania and then at Dartmouth College in New Hampshire. While at Dartmouth,
he was also president of the American Psychological Association. Early in his career,
Köhler applied gestalt psychology to auditory perception but then expanded gestalt
psychology into other domains of psychology, including his famous study of problem
solving in chimpanzees.
Direct Perception (The Gibsonian Approach)
The direct perception view was developed by the American husband-and-wife team
J. J. and Eleanor Gibson, who were professors at Smith College in Northampton,
Massachusetts, and later at Cornell University in Ithaca, New York (Gibson, 2001). The
Gibsons emphasized that the information in the sensory world is complex and abun-
dant, and therefore the perceptual systems need only directly perceive such complexity.
In this view, the senses do not send to the brain incomplete and inaccurate informa-
tion about the world that needs to be reasoned about to generate a perception. Thus,
the direct perception view is diametrically opposed to Helmholtz’s concept of uncon-
scious inference. Rather, in the direct perception view (Gibsonian approach), the world
generates rich sources of information that the senses merely need to pick up directly.
The direct perception view also emphasized ecological realism in experiments. This
means that rather than showing simple displays to participants in experiments, direct
perception theorists advocated using more naturalistic stimuli. Indeed, J. J. Gibson
(1979) criticized much work in perception because the stimuli used by researchers were
just points of light, tones, or stimuli that otherwise would not normally be encoun-
tered in everyday life. He emphasized that researchers should study real-world stimuli.
For this reason, the direct perception view is often called the ecological approach to
perception. For a demonstration of optic flow, one of the key contributions of this
approach, see ISLE 1.9.
Information-Processing Approach
The information-processing approach postulates that perceptual and cognitive sys-
tems can be viewed as the flow of information from one process to another. Information
is collected by sensory processes and then flows to a variety of modules that decode the
information, interpret it, and then allow the organism to act on it. Consider the pro-
cessing of visual information. Information flows from the eyes to various units in the
brain that extract color, motion, figure–ground, and object information and then pass
that information to cognitive systems that extract meaning and then pass it to other
cognitive systems that determine actions that should be implemented on the basis of
the visual information. The information-processing view stipulates that each of these
stages or processes takes a finite amount of time, even if they are very fast, and therefore
these processes can be observed or measured by recording reaction times as observers
do various tasks (Figure 1.19).
The information-processing view greatly influenced cognitive psychology. During
the 1960s and 1970s, information-processing models were used in both perception
research and memory research. The approach continues to influence both of these
fields, though contemporary researchers realize that the brain is massively “parallel”
and many processes can be occurring simultaneously.
FIGURE 1.18
Wolfgang Köhler
Wolfgang Köhler (1887–1967) was
a major theorist in the gestalt
psychology movement. Köhler
immigrated to the United States
prior to World War II. He expanded
the use of gestalt psychology
beyond its traditional domain of
sensation and perception to other
areas of psychology, such as
problem solving.
©
IN
TERFO
TO
/A
lam
y
ISLE 1.9
Optic Flow
Direct perception
(Gibsonian approach): the
approach to perception that
claims that information in the
sensory world is complex
and abundant, and therefore
the perceptual systems need
only directly perceive such
complexity
Ecological approach to
perception: another name for
the direct perception view
Information-processing
approach: the view that
perceptual and cognitive
systems can be viewed as the
flow of information from one
process to another

16 Sensation and Perception
The information-processing
view is different from Gibson’s
direct perception view because,
like Helmholtz’s view, the informa-
tion-processing view requires inter-
nal cognitive processes to interpret
the perceptual image, whereas the
direct perception view asserts that
the sensory input is sufficient in
and of itself. The information-pro-
cessing view is similar to the gestalt
view in the specification of internal
processes that extract information.
However, the gestalt view emphasizes
patterns and organization, whereas the
information-processing view empha-
sizes the analysis of information and
its flow from one system to another.
Computational Approach
The computational approach stud-
ies perception by trying to specify
the necessary computations the
brain would need to carry out to
perceive the world. Originally devel-
oped by David Marr (1982), it was
heavily influenced by the growth of
computer science and, in particular,
early theory in artificial intelligence.
Marr attempted to specify perception
in terms of what computations the brain would need to perform the task of perception.
He conceived of the brain as an incredibly complicated computer and sought a math-
ematical explanation for perceptual processes, especially vision. The computational
approach built on the information-processing approach but acknowledged from the
beginning that many perceptual processes may occur in parallel in the brain. Using the
approach’s modern form, researchers attempt to develop mathematical models that pre-
dict perceptual phenomena. Many of these mathematical models are based on neural
networks, computer simulations of how nervous systems work (Venrooij et al., 2013).
Many who value the computational approach try to get computers to “see” in
ways that make sense given our knowledge of the brain. In this way, the computa-
tional approach is often more theoretical, focusing on modeling perception in com-
puter simulations, whereas the information-processing approach is more directly
linked to observations in behavioral experiments. In computational models, if the
computer can “see,” theoretically so should the brain, and it is quite possible that the
brain will also use many of the same computations as the computer. Studies using
this approach often attempt to simulate perception on computers. A researcher will
give a computer some visual task. The task may be to see a real object, and the com-
puter has some form of electronic “eye,” such as a video camera, attached. It might
also involve simply a computer-generated world. In either case, the computer may
be given the task of identifying an object, perhaps partially hidden by another object
(e.g., Doerschner, Kersten, & Schrater, 2009). If the computer can do the task and
also tends to make the same types of mistakes humans do, then those doing research
FIGURE 1.19 An Information-Processing Model
Information processing means that information flows from one process or stage to
another during perception. Thus, in the diagram, we see an early sensory stage in which
transduction takes place. Information then flows to a series of perceptual modules,
eventually leading to final percept, which drives action.
Visual Image Storage
Long-Term Memory
Cognitive Processor
Movement Response
(arms, legs, mouth, eyes, etc.)
Motor
Subsystem
Cognitive
Subsystem
Perceptual
Subsystem
Senses
Perceptual Processor
Auditory Image Storage
Working Memory
Motor Processor
Computational approach:
an approach to the study
of perception in which the
necessary computations the
brain would need to carry out to
perceive the world are specified

17 Chapter 1: What Is Perception?
according to the computational approach have some
confidence that they have made progress in their goal.
Neuroscience in Sensation
and Perception
The goal of neuroscience is to understand sensation and
perception in terms of the structures and processes in
the nervous system that produce it. Thus, the neurosci-
ence approach starts with examining the physiological
processes whereby a physical signal is transduced into a
neural signal. Neuroscience then continues to investigate
sensation and perception by looking at connections from
the sensory organs to the brain and then at regions in
the brain itself that are involved in perceptual processes.
Neuroscience research can span the gamut from work
on single cells within the brain to examining connections
among differing regions in the brain.
Neuroscience is interested in the cellular level, nec-
essary to understand how individual neurons convert
physical stimuli into electrochemical signals (Figure 1.20). At the cellular level, neu-
roscientists can look at the actions of individual cells and how they respond to
particular signals. Neuroscience is also interested in what processes occur in the
brain to process and interpret sensory information. Here, neuroscientists can look at
larger units in the brain and attempt to correlate those regions with particular per-
ceptual functions. For example, studies show activity in an area of the brain, known
as MT (middle temporal) or V5, in the occipital lobe, when people are watching
moving stimuli.
One of the most important developments in neuroscience was the development of
the microelectrode in the 1940s and 1950s. A microelectrode is a device so small that
it can penetrate a single neuron in the mammalian central nervous system without
destroying the cell. Once in the cell, it can record the electrical activity there or even
stimulate the cell by carrying electrical current to the cell from an electrical source at
the command of the scientist. Thus, this method allows the recording of the behavior
of single neurons in the mammalian brain. It was first used in the sensory systems
by Kuffler (1953) (ISLE 1.10). In this way, it can be used to determine what kind of
stimuli a particular cell responds to. Thus, for example, a cell in area V1 (also known
as the primary visual cortex) of the brain may respond to visually seen lines of dif-
ferent orientations. Another cell might respond only to stimuli in the left visual field.
This technique led to some profound breakthroughs in our understanding of how the
brain processes sensory information.
Hubel and Wiesel (1959, 1962, 1965) are probably the names most associated
with this technique. Their Nobel Prize–winning work not only helped us understand
the behavior of individual cells but uncovered unexpected levels of organization in
the brain (Hubel & Wiesel, 1965) and information on how the brain develops (e.g.,
Hubel & Wiesel, 1962). For example, they found cells in the brains of cats and
monkeys that were selective to one eye or the other and other cells that responded
only when both eyes were able to see the same objects in the visual world. Single-
cell recording made major contributions to neuroscience with animal models, but it
is used less often nowadays because it is difficult to do with humans and we have
increased concerns about animal welfare.
ISLE 1.10
Kuffler and Single-Cell Recording
FIGURE 1.20 A Sensory Neuron
Bipolar cells, such as the one depicted in this image, connect the
receptor cells to the tracts that lead information from the sensory
organ to the brain. Bipolar cells in vision, for example, connect
rods and cones to the cells that lead information out of the eye and
through the optic nerve.
©
The Science Picture Com
pany/A
lam
y
Neuroscience: the study of the
structures and processes in the
nervous system and brain
Microelectrode: a device so
small that it can penetrate a
single neuron in the mammalian
central nervous system without
destroying the cell

18 Sensation and Perception
The neuroscience approach also includes the field of neuropsychology.
Neuropsychology is the study of the relation of brain damage to changes in behavior.
In neuropsychological research, scientists know what part of a person’s brain is dam-
aged and then look for behavioral changes in that individual. Behavioral changes
may include problems with language, memory, or perception. They may also include
difficulties in decision making, action, or emotion. Brain damage may arise from
a variety of tragedies and accidents, including strokes, tumors, aneurysms, loss of
blood during surgery, auto accidents, blows to the head, bullet wounds, concus-
sions, and near drowning. The origins of neuropsychology arise from such tragedies.
Indeed, the study of neuropsychology began in earnest in the 1870s in Europe, as
antiseptic procedures allowed soldiers in the Franco-Prussian war to survive gun-
shot wounds that would previously have been fatal. Antiseptics arrived too late for
American soldiers in the Civil War, many of whom died from infections that would
have been curable a few years later. Neuropsychologists continue to learn from and
help individuals returning from war (Vasterling et al., 2006).
Consider a particular case in which brain damage resulted from lifesaving sur-
gery. A patient called D.B. had part of his occipital lobe removed as treatment for
a malignant tumor. The tumor was causing seizures and could have threatened his
life, were it to have spread to other regions of his brain and body. Thus, surgeons
removed it as carefully as possible but also had to excise a great deal of tissue within
the V1 area of his occipital lobe. Medically, the operation was a tremendous success.
The tumor removal allowed D.B. to return to his job as a computer programmer and
lead a normal life, with few or no side effects. However, as a result of the loss of area
V1 tissue, D.B. developed partial blindness in one area of his visual field (Weiskrantz,
1996). This sounds more serious than it actually was. D.B. had no difficulties mov-
ing his eyes, so his eye movements quickly compensated for this patch of blindness.
However, it allowed D.B. to be an interesting research participant. The researchers
could correlate the brain damage (known to be in a particular area of the occipi-
tal lobe) and the behavioral deficits (partial blindness). As such, it was possible to
correlate the area of the occipital lobe that was removed with being responsible for
seeing in a certain part of the visual field. D.B. was also an important patient in the
study of the phenomenon of blindsight, a topic that is covered in depth in Chapter 4.
One form of brain damage affecting perception is known as agnosia. Agnosia is
a deficit in some aspect of perception as a result of brain damage. For example, there
is a form of agnosia caused by damage to an area of the temporal lobe known as the
fusiform face area (prosopagnosia, or face agnosia). Damage to this area results in
a deficit in recognizing faces. Damage to certain areas of the right temporal lobe can
result in a person’s loss of appreciation of music. This condition, known as amusia,
a form of agnosia, is covered in Chapter 13. Thus, neuropsychological methods are
effective at relating areas of the brain to particular perceptual functions.
Neuroscience research also includes the neuroimaging techniques. Neuroimaging
involves technologies that allow us to map living intact brains as they engage in
ongoing tasks. Neuroimaging allows us to observe an intact living brain as it per-
ceives, learns, and thinks. This is the newest technology used in neuroscience, with
the first magnetic resonance imaging (MRI) research not beginning until the 1990s.
Since then, however, neuroimaging research has become a major force in under-
standing both the brain and cognition and perception. One method is functional
MRI (fMRI). This technique can image the blood levels in different areas of the brain,
which correlate with activity levels in those regions, allowing activity in the human
brain to be correlated with our actual sensory abilities (Tremblay, Dick, & Small,
2011). Figure 1.21 shows what an fMRI scan looks like.
The fMRI technique is thought of as a hemodynamic technique because it mea-
sures the blood flow to the brain. The technique starts with the reasonable assumption
Neuropsychology: the study
of the relation of brain damage
to changes in behavior
Agnosia: a deficit in some
aspect of perception as a result
of brain damage
Prosopagnosia: face
agnosia, resulting in a deficit in
perceiving faces
Amusia: a condition in which
brain damage interferes with
the perception of music but
does not interfere with other
aspects of auditory processing
Neuroimaging: technologies
that allow us to map living
intact brains as they engage in
ongoing tasks
Functional magnetic
resonance imaging (fMRI):
a neuroimaging technique that
generates an image of the brain
on the basis of the blood levels
in different areas of the brain,
which correlate with activity
levels in those regions

19 Chapter 1: What Is Perception?
that because the brain is a biological organ, it requires oxy-
gen. Moreover, the areas of the brain that are active require
more oxygen and therefore more blood than other areas of
the brain. Thus, a person talking will need oxygen delivered
to the parts of his or her brain that are responsible for speech
production. A person listening to music will require a greater
oxygen supply to areas that are responsible for music percep-
tion. Thus, if you can trace the flow of blood in the brain, you
will know what areas of the brain are currently in use. For
example, Kau et al. (2013) recorded the neural activity in the
brains of participants tracking moving dots across a screen.
When the participants were attending to the motion, there
was activity in an area of the brain known as V3, located in
the occipital lobe. Other studies have associated area V3 with
movement perception.
Neuroimaging may very well be the fastest growing area in
scientific psychology. It offers us the opportunity to watch the
human brain as it does its work, certainly a fascinating propo-
sition. In this way, it can tell us a great deal about the structure
and function of the human brain and the nature of sensation
and perception. But neuroimaging can also be misleading. For
example, small differences in methodology between studies
can sometimes lead to big differences in the neural patterns.
Moreover, because fMRI researchers may do many statistical
tests per scan, there is a risk that neural activity in some regions may exceed a cri-
terion level simply by chance (Bennett, Wolford, & Miller, 2009). It is also true that
neuroimaging, at present, is still looking at vast groupings of cells. So neuroimaging
technology still cannot bridge the gap from gross anatomy to the fine-tuned results
of single-cell studies. Thus, it is still worthwhile to cast a critical eye at data that
come from fMRI studies, as they are not immune to such methodological problems.
Nonetheless, neuroimaging is a useful tool and one that has definitely broadened the
scope of the neuroscience of perception.
TEST YOUR KNOWLEDGE
1. What was at the center of the difference in viewpoints of Helmholtz and Hering?
Does this difference in view still have resonance for today’s science?
2. What is the difference between the views of gestalt psychology, direct
perception, the information-processing view, and the computational approach?
1.5
Explain the theory of cognitive penetration, and apply
sensation and perception research to collisions.
FIGURE 1.21 fMRI
This functional magnetic resonance image shows
the areas of the brain that are active when an odor is
presented to an individual. The areas depicted in color
include areas of the brain critical in odor perception, such
as the piriform cortex in the temporal lobe.
©
BSIP SA
/A
lam
y
EXPLORATION: Cognitive Penetration
Basketball players will sometimes say that the basket-
ball rim looks wider when they are doing well and looks
narrower when they are not making their shots. Golfers
report similar experiences with their putting. Tired hikers
may say the mountain looks steeper when they are tired
than when they are fresh. Many of us will report colors
being brighter when we are in good moods than when
we are depressed. Most of us agree that food tastes better
when we are hungry. Is this really true? Do factors such
as our success, our physical state, and our emotional state

20 Sensation and Perception
really affect how we perceive? Does our state of hunger
really affect what our taste buds are reporting?
The view that cognitive and emotional factors influence
the phenomenology of perception is known as cognitive
penetration (Marchi & Newen, 2015). Cognitive pen-
etration means that nonperceptual factors affect what
we see, hear, taste, and feel. The opposing view is that
perception is not affected by cognitive factors and that
only our reporting of perception is. This view is called
cognitive impenetrability (Firestone & Scholl, 2016).
Impenetrability implies that our perception remains the
same, regardless of our cognitive and emotional state.
What changes instead is attention, expectation, or our
mood state, which is different than our perceptual state.
The dominant view in the field is that perception is cog-
nitive impenetrable. However, recent research supports
some instances in which cognitive or emotional factors
influence the phenomenology of the perceived world.
Delk and Fillenbaum (1965) asked participants to match
the color of figures with the color of their background.
Some of the figures depicted objects associated with a par-
ticular color. These included typically red objects such as
an apple, lips, and a symbolic heart. Other objects were
presented that are not typically associated with red, such
as a mushroom or a bell. However, all the figures were
made out of the same red–orange cardboard. Participants
then had to match the figure to a background varying from
dark to light red. They had to make the background color
the same as the color of the figures. Delk and Fillenbaum
found that red-associated objects (e.g., the apple) required
more red in the background to be judged a match than
did the objects that were not associated with the color red.
This suggests that the knowledge of the objects was influ-
encing people to perceive them as being more red than
other objects (Figure 1.22). Hansen, Olkkonen, Walter,
and Gegenfurtner (2006) replicated this basic finding.
They presented participants with photographs of fruit
such as bananas and simple patches of colors. Participants
were asked to adjust the color of the object (banana or
patch) to the uniform gray background. Hansen et al.
found that participants adjusted the fruits differently than
the patches of equivalent color. For example, participants
added more blue to the banana in order to cancel out the
more perceived yellowness of it (we will discuss color
cancellation more in Chapter 6). Hansen and colleagues
suggest that this implies that the cognitive association of
objects to color influences how we perceive that color.
There may be cognitive penetration in other sensory
domains as well. For example, in research on pain, there
have been many studies documenting the placebo effect,
which is the finding that when people expect reduced
pain from a medicine, they experience less pain, even if
the medicine is essentially inert (Miller & Miller, 2015).
Freeman and colleagues (2015) gave three identical
creams to three different groups of participants before
presenting a somewhat painful stimulus. In the first con-
dition, they told participants that the cream was “lido-
caine” and would reduce pain (i.e., an analgesic). In the
second condition, they told participants that the cream
was “capsaicin” and might increase the pain (i.e., hyper-
algesia). In the third condition, participants were told
the cream was “neutral”; that is, it would not affect their
pain levels at all. The belief that the cream was an anal-
gesic led to less pain experienced relative to the control
condition. The belief that the cream was hyperalgesic
(leading to greater pain) led to more reported pain rela-
tive to the control. These results suggest that our expec-
tations actually influence the pain we experience (but see
Gatzia & Schwartz, 2016).
Despite the evidence, cognitive penetration is a tricky
issue. There are lots of alternate explanations for why
the results might be the way they are instead of the
Cognitive penetration: the view that cognitive and
emotional factors influence the phenomenology of perception
Cognitive impenetrability: perception is not affected by
cognitive factors; only our reporting of perception is
FIGURE 1.22 Cognitive Penetration Experiment
(Delk & Fillenbaum, 1965)
Red-associated objects (e.g., the apple) required more red in the
background to be judged a match than did the objects that were
not associated with the color red.
©
iS
to
ck
ph
ot
o.
co
m
/A
le
xB
ry
lo
v

21 Chapter 1: What Is Perception?
view that perceptual phenomenology actually changes
(Firestone & Scholl, 2016). For example, in one study,
participants judged a hill to be steeper when they were
wearing heavy backpacks (Bhalla & Proffitt, 1999).
Firestone and Scholl point out that many factors other
than perception may go into such judgments, such
as the expectation that a mountain must be climbed.
When participants had to match the hill they saw with
a test incline, they did not show different choices based
on the heaviness of the backpack. Thus, Firestone and
Scholl think that nonperceptual factors influenced their
decisions, not their perceptions. Indeed, the idea of cog-
nitive penetration is now one generating much interest
and controversy in the field of perception. Keep an eye
on this field, as we suspect there will be some very inter-
esting findings to come yet ahead.
APPLICATION: Avoiding Collisions
There is probably no field of psychology in which there
have been greater applications of psychological work than
the area of sensation and perception. It is also likely that
applications of this research area will continue to grow in
the future. We use our perceptions of the world around us
in nearly everything we do in our lives, with the possible
exception of daydreaming. Every practical aspect of life
involves looking, listening, touching, and tasting. Thus,
if you are considering a career in applied psychology or
human factors, you should pay very close attention to the
topics in this book and to your course in sensation and
perception. Understanding the basics of sensation and
perception is critical for any human factors psychologist
working in industry or government. Understanding the
basics of sensation and perception is important for civil
engineers who are designing safer roads.
Just consider a few of our technological horizons. Think
of designing the face of a modern smartphone. The size
of text and icons must be large enough for the human eye
to distinguish, and buttons must be positioned so that
human fingers can distinguish between the button for the
phone and the button for the music player. Companies
such as Apple and Samsung hire many human factors
psychologists to help them design these products in ways
that are consistent with human perception. Moreover,
consider the console of a modern jet plane. Warning
lights and controls must be positioned so that they
quickly demand the attention of the pilot if there is a
problem with a particular system in the airplane.
Sensory systems are also important in developing new
technologies. A major technological innovation that
many car companies are working toward is functional
self-driving cars. What goes into engineering such a
system? Cameras and computers have to be wired like
eyes and brains (only without blinking, falling asleep,
or experiencing road rage). Understanding the cues the
human eye attends to while driving is critical in design-
ing safe self-driving cars (Brett, 2016). Consider also
computerized voice recognition systems. The engineers
who design these systems must first understand the pro-
cesses by which humans recognize one another’s voices
and interpret meaning. Thus, several of our major tech-
nological fronts involve a role for understanding human
perception.
Returning to cars, it may be many years before self-driv-
ing cars are widespread, but road safety is a major issue
both in the United States and throughout the world.
According to U.S. government statistics, there are more
than 32,000 fatalities per year in car accidents in the
United States alone. This translates into more than
2,600 deaths per month on American roads. That means
that approximately the same number of Americans die
in car accidents every month as did in the tragic attacks
of September 11, 2001. It is estimated that there were
5.5 million automobile accidents in the United States in
2010 (Keane, 2011). Thus, there is still much that can
be done to improve road safety outcomes. Sensation and
perception research has much to offer here.
Many sensation and perception researchers study the
visually guided action associated with driving cars or
flying planes. Some of this research concerns the effects
of distraction on driving, a topic we discuss in depth
later (Chapter 5). Other research considers the effect
of fatigue on perception and driving ability. In this
Application section, we consider the perceptual pro-
cesses that drivers use to avoid automobile accidents.
Think about driving (assuming that you are a driver).
It is a visual task: We look in front of us to see where
we are going, and we check our rearview mirrors often
for potential dangers coming from behind us. We use
this visual information for simple tasks, such as noticing

22 Sensation and Perception
when our exit is coming and moving to the right or seeing
a swerving car and quickly applying the brakes. In this
way, avoiding impending collisions involves perceptually
guided action. When people see approaching objects, their
visual systems rely on their perception of depth, that is,
how far away the oncoming objects are, and judgments of
time to collision. These judgments need to be very accu-
rate in order to make good driving decisions.
However, research shows that people have systematic
biases in making these decisions. People estimate that
a large but farther away approaching object will collide
with them sooner than a smaller but closer approach-
ing object (DeLucia, 2013; DeLucia et al., 2016). That
is, people estimate accurately the likelihood that a truck
will strike them but underestimate the likelihood that
a smaller object such as a motorcycle will enter their
path. Motorcyclists often complain about how cars “cut
them off.” This may not always be the result of reck-
lessness or rudeness but rather a poor estimate on the
part of the car’s driver of how close the motorcycle
actually is (Figure 1.23). We now consider the research
of Dr. Patricia DeLucia of Texas Tech University, who
has devoted her career to understanding the factors that
affect our perceptual decisions in collision situations
(DeLucia, 2013; DeLucia et al., 2016).
Why is this? DeLucia’s research shows that the human
visual system uses two cues to make judgments about
impending collisions. The first is called time to collision.
This means that, in theory, your visual system estimates
the time it will take for an object to collide with you
by dividing the object’s optical size at a given point in
time by the object’s rate of expansion within the visual
field. That is, time to collision is determined by the
object’s optical size per unit time. In simpler language,
approaching objects increase in size, and more rap-
idly approaching objects increase in size more as they
approach you. Think of a baseball player judging when
to swing his bat. A fastball increases in optical size as it
approaches him more so than does a changeup (a pitch
thrown deliberately slow). Or think of a pilot guid-
ing a plane to touchdown on a runway. As the plane
approaches the ground, the runway increases in optical
size for the pilot. There is ample evidence that people
attend to this aspect of looming objects in a variety of
situations (Hecht & Savelsbergh, 2004).
However, people also use another cue to determine
time to collision. The second cue is the size of the
object. Larger objects are judged to be closer than
smaller objects. This is known as the size–arrival effect
(DeLucia, 2013). Numerous studies have shown that
perception of collision is
affected by the relative
sizes of objects (Brendel,
DeLucia, Hecht, Stacy, &
Larsen, 2012; DeLucia, 2013; Hahnel & Hecht, 2012).
You can see an illustration of this effect in ISLE 1.11.
The size–arrival effect results in the illusion that smaller
objects are less likely to collide with the viewer. This has
a number of unfortunate consequences for driving and
transportation safety. For example, drivers may under-
estimate the likelihood of collision when turning when
a smaller oncoming vehicle is approaching because it is
perceived as being farther away. Indeed, data from both
actual accidents and experimental simulations show
that a crash between a motorcycle and a car can occur
when the car intrudes on the motorcycle’s path, with
the car’s driver thinking the motorcycle is farther away
(Brendel et al., 2012; DeLucia, 2013; DeLucia et al.,
2016). In one simulated study, participants were asked
to press a button when they thought an approaching
vehicle would arrive at a particular location (Horswill,
Helman, Ardiles, & Wann, 2005). The experiment-
ers varied the sizes of the vehicles (motorcycle, car, or
van) and their speeds (30 or 40 mph). Time-to-collision
estimates were greater for motorcycles than for cars
and vans, consistent with the size–arrival effect. Thus,
DeLucia and others have recommended that roads and
motorcycle design should consider this factor in order
to make transportation safer. It is also possible that
ISLE 1.11 Size–Arrival Effect
©
iS
to
ck
ph
ot
o.
co
m
/p
al
in
ch
ak
jr
FIGURE 1.23 Avoiding Collisions
The automobile driver must estimate the time to collision with the
bicyclist. If the driver determines that a collision is imminent, he
or she must apply the brakes immediately, in this case, to spare
the bicyclist major injury.
Time to collision: the estimate of the time it will take for an
approaching object to contact another
Size–arrival effect: bigger approaching objects are seen
as being more likely to collide with the viewer than smaller
approaching objects

23 Chapter 1: What Is Perception?
education can be directed at drivers to help them realize
that they often underestimate the likelihood of collision
with smaller objects. Motorcycles designed to look big-
ger may also be safer for motorcyclists to ride. Thus,
this first Application section introduces the concept of
applications of sensation and perception research and
describes a particular area, research on time to collision
and its implications.
CHAPTER SUMMARY
1.1
Discuss why understanding sensation and per-
ception is important.
The study of sensation and perception sheds light on the
basic nature of what it is to be human. Sensation is the
registration of physical stimuli on sensory receptors, and
perception is the process of creating conscious perceptual
experience from sensory input. In this textbook, we discuss
the science and applications of research on sensation and
perception. In this chapter, we have discussed the nature of
physical stimuli, whether they be light for vision on the retina
or sound waves for the auditory system.
1.2
Assess why there are actually more than five
senses.
In addition to vision, hearing, touch, smell, and taste, we
have a vestibular system to help keep our balance and a pro-
prioception system to allow us to monitor the position of our
bodies. Our sense of touch is composed of multiple systems
designed to sense different features of the environment. Heat,
coldness, pain, itchiness, and soft touch are all implemented
by separable sensory systems.
1.3
Describe how transduction transforms a physi-
cal signal into a neural signal.
Sensory systems transduce physical signals into neural
responses, which are sent to the brain for processing. The
brain processes the signals, determines their meaning, and
decides on appropriate actions. Perception also produces a
characteristic phenomenology, which is the purely subjec-
tive experience we get when perceiving the world.
1.4
Illustrate the history of the study of sensation
and perception.
Writings on disorders of sensation and perception go back all
the way to the ancient Egyptians. Aristotle theorized extensively
about perception and its causes. Later, in the 19th century, German
physiologists began experimenting on the neural processes that
underlie sensation, and others started the field of psychophysics,
which studies the relation of physical stimuli to the psychologi-
cal experience. Later influences in the development of sensation
and perception research include gestalt psychology, Gibsonian
direct perception, information processing, and the computational
approach. Neuroscience also addresses issues of sensation and
perception. Neuroscience research includes single-cell record-
ing, neuropsychology, and neuroimaging.
1.5
Explain the theory of cognitive penetration, and apply
sensation and perception research to collisions.
Cognitive penetration means that nonperceptual factors affect
what we see, hear, taste, and feel. The opposing view is that
perception is not affected by cognitive factors and that only
our reporting of the perception is. This view is called cognitive
impenetrability. Impenetrability implies that our perception
remains the same, regardless of our cognitive and emotional
state. In the last section, we took one topic in depth as a case
study in applications of sensation and perception research.
This research concerns the visual perception of imminent col-
lisions. The research shows that people use both the rates of
expansion of approaching objects and the sizes of the objects
to estimate time to collision. The use of the sizes of objects
to estimate time to collision results in errors, in that smaller
objects are judged to impact later than they actually do.
REVIEW QUESTIONS
1. What is the myth of five senses? Can you list eight
different human sensory systems?
2. What do the terms sensation and perception mean?
What is the difference between the two?
3. What is the process of transduction? Why is it
important to perception?
4. What is phenomenology? Why is it so difficult to
address in science?

Sensation and Perception24
5. Who was Hermann von Helmholtz? What was his view of
color vision? How did it differ from that of Ewald Hering?
6. What is an unconscious inference in perception?
Why is it important to the constructivist approach?
7. What is a JND? Can you give a real-world example of
a JND?
8. What is the direct perception view? How does it dif-
fer from the information-processing view?
9. What is cognitive impenetrability? How does it differ
from cognitive penetration?
10. What is meant by the term time to collision? How is it
that the size–arrival effect changes our judgments of
collision times? What practical suggestions could you
make to improve transportation safety on the basis of
these findings?
PONDER FURTHER
1. Our perceptual abilities are shaped by natural evo-
lution. Weber’s law states that a just-noticeable
difference between two stimuli is related to the
magnitude or strength of the stimuli. Why might it
be adaptive to have a system that has this kind of
varied sensitivity as a function of the magnitude of
the stimuli? That is, why can we tell the difference
in weight between a 1-gram weight and 2-gram
weight, but not between a 2,000-gram weight and a
2,001-gram weight?
2. Most likely, you have a piece of art on your wall.
Maybe it is something original, perhaps a poster of a
famous work of art. Look at it now. Can you see if the
artist applied any gestalt principles in constructing the
visual image? Does knowing about the gestalt princi-
ples change the way you appreciate the art?
KEY TERMS
Action, 8
Aftereffect, 10
Agnosia, 18
Amusia, 18
Cognitive impenetrability, 20
Cognitive penetration, 20
Computational approach, 16
Constructivist approach, 11
Direct perception
(Gibsonian approach), 15
Doctrine of specific nerve energies, 11
Ecological approach to perception, 15
Functional magnetic resonance
imaging (fMRI), 18
Gestalt psychology, 13
Information-processing approach, 15
Microelectrode, 17
Neural response, 6
Neuroimaging, 18
Neuropsychology, 18
Neuroscience, 17
Perception, 5
Phenomenology, 8
Prosopagnosia, 18
Psychophysics, 13
Receptors, 6
Sensation, 5
Size–arrival effect, 22
Stimulus, 5
Time to collision, 22
Transduction, 6
Unconscious inference, 11
Weber’s law, 12

Chapter 1: What Is Perception? 25
Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e
SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review,
study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning.
Learning Objectives Digital Resources
1.1 Discuss why understanding sensation and perception is
important.
Wishful Seeing: How Preferences Shape Visual Perception
David Eagleman: Can We Create New Senses for Humans?
A Virtual Arm That Talks With the Brain
1.2 Assess why there are actually more than five senses. We Have Far More Than Five Senses
1.3 Describe how transduction transforms a physical signal into a
neural signal.
Perception and Health
Bright Lights, Big Headache: A Study Explains
1.4 Illustrate the history of the study of sensation and perception. Gestalt Illusion
Waterfall Illusion
Artwork of Bev Doolittle
1.5 Explain the theory of cognitive penetration, and apply sensation
and perception research to collisions.
Effects of Size on Collision Perception and Implications for
Perceptual Theory and Transportation Safety
Box Shape Influences the Size-Weight Illusion During
Individual and Team Lifting

Research Methodology
Jon Berkeley/Ikon Images/Getty Images
LEARNING OBJECTIVES
2.1 Explain the nature of psychophysical scales and how they measure the
relations between real-world stimuli and our perceptions of them.
2.2 Demonstrate an understanding of signal detection theory.
2.3 Evaluate neuroscience methods and what they tell
us about sensation and perception.
INTRODUCTION
In 2016, the Trinidad Moruga Scorpion blend was ranked the world’s second hottest
chili pepper after the Carolina Reaper (Bannister, 2016). According to the Chile Pepper
Institute at New Mexico State University, the Trinidad Moruga Scorpion rates close to
3 million on the Scoville scale. That puts this pepper in roughly the same range on the
Scoville scale as law enforcement–grade pepper spray. By contrast, Scotch bonnet peppers
and habanero peppers score from 100,000 to 300,000 Scoville units. Scotch bonnets and
habaneros are usually the hottest peppers most people eat. The first question we can ask is
why would anyone ever want to eat a Trinidad Moruga Scorpion? And the next question
we can ask is how do we determine hotness and how people experience it?
The Scoville scale measures our detection of the amount of an ingredient called
capsaicin in chili peppers. Capsaicin is a chemical present in peppers that directly
stimulates the somatosensory system, especially heat and pain receptors in our
mouths, and our eyes (which is why pepper spray irritates the eyes). It presumably
evolved to be present in wild chili peppers to prevent mammals from consuming their
fruit. Capsaicin, in general, repels mammals but does not affect birds, as the birds are
necessary to disperse the seeds. However, many people acquire a taste for capsaicin,
as the heat and pain felt in the mouth interact with smell and taste to create inter-
esting, if acquired, flavors, known as piquancy. Indeed, worldwide, the chili pepper
has become an important ingredient in cooking. Moreover, some chili pepper enthu-
siasts “compete” in competitions to see who can eat the hottest pepper. For example,
there is a chili pepper–eating contest at the North Carolina Hot Sauce Contest. To
win, contestants must eat an entire orange habanero chili (300,000 Scoville units)
(Figure 2.1). In the competition, contestants start off with milder peppers and work
their way up to the really spicy stuff. To find out more about this competition, see
the link on ISLE 2.1.
ISLE EXERCISES
2.1 North Carolina Hot
Sauce Contest Link
2.2 Method of Limits
2.3 Method of Constant
Stimuli
2.4 Method of
Adjustment
2.5 Point of Subjective
Equality
2.6 Magnitude Estimation
2.7 Stevens’s Power Law
2.8 Forced-Choice Method
2.9 Signal Detection
Experiment
2.10 Signal Detection
Theory
2.11 Signal Detection
Theory and the Receiver-
Operating Characteristic
(ROC) Curve
2.12 Masking
Demonstration
2.13 Seeing With
Myopia and Presbyopia
2
Scoville scale: a measure of
our detection of the amount of
an ingredient called capsaicin in
chili peppers
Capsaicin: the active
ingredient in chili peppers that
provides the experience of
hotness, piquancy, or spiciness

28 Sensation and Perception
The Scoville scale is a psychophysical scale, meaning that it relates psychological
experience to some aspect of the physical world. A psychophysical scale is one in
which people rate their psychological experiences as a function of the level of a phys-
ical stimulus. In this case, the Scoville scale measures our experience of piquancy or
“hotness” at different concentrations of capsaicin. Thus, the concentration of cap-
saicin in a particular food or drink is the physical dimension, whereas our experi-
ence of piquancy is the psychological correlate of that physical dimension. We will
encounter numerous instances of psychophysical scales throughout this textbook.
For example, the decibel scale measures the psychological construct of loudness as a
function of the intensity of a sound stimulus.
The Scoville scale is named after a chemist named Wilbur Scoville, who developed
the scale more than 100 years ago, in 1912. He derived the scale before chemists
could accurately measure the amount of capsaicin in a particular pepper, which labs
can do routinely now. In Scoville’s time, having a reliable scale to measure the “hot-
ness” of foods was important in the development of commercial foods. To measure
the piquancy, Scoville recruited brave volunteers from the company he worked for to
evaluate chili peppers. Once he had willing volunteers, Scoville adjusted the level of
chili peppers in a liquid solution that he had his observers taste. Obviously, the greater
the concentration of chili peppers, the hotter would be the taster’s experience, but
Scoville wanted to quantify this relation. He measured when the observers first start-
ing experiencing heat when tasting the pepper solution and how much more capsaicin
he needed to add to get people to notice a difference in the hotness of the solution. If a
taster did not notice any difference as a function of more chili peppers, Scoville added
more until that taster did notice the difference. In this way, Scoville was measuring a
characteristic known in psychophysics as the just-noticeable difference (JND). A JND
is the smallest amount of physical change observers notice as a perceptual change.
The Scoville scale, then, measures people’s perceived sense of hotness as a func-
tion of the amount of capsaicin in the sample solution. This scale was critical for
the development of commercial products that contained capsaicin. The scale is now
widely used in the food industry. The production of commercial hot sauces is a world-
wide business, and companies will report the Scoville ratings of their sauces. Tabasco
sauce, spicy to many of us, rates only about 2,000 on the Scoville scale, far less than
the piquancy of a Scotch bonnet pepper, let alone a Trinidad Moruga Scorpion. So, if
you know that Tabasco sauce is relatively mild on the Scoville scale, you might try it
on your tofu. However, if someone is offering you Trinidad Moruga Scorpion sauce,
you might want to think twice before dousing your dinner. For our purposes here,
the perception of piquancy and the Scoville scale serve as an excellent introduction
to the research methods used in sensation and perception and in understanding the
concepts of psychophysics.
This chapter covers basic questions, techniques, and research methods used by
researchers to discover how human sensory systems work. This information will set
the stage for later chapters that cover specific sensory systems, how they work, and
what we perceive. In terms of methods, in some cases, participants are asked to rate
a stimulus on a numerical scale, whereas in other situations, they may be asked to
decide only whether a stimulus is present (see the various demonstrations on ISLE).
In the latter case, a researcher can quantify on the basis of the percentage of correct
responses at each level of the stimulus.
The study of human sensory systems starts with psychophysics, the study of the
relation between physical stimuli and perception events, much as we just saw with
the Scoville scale for chili pepper hotness. Psychophysical methods involve presenting
a carefully controlled stimulus to a participant and asking a question directly of the
participant that allows the answer to be quantified, that is, turned into a number. From
©
iS
to
ck
ph
ot
o.
co
m
/D
N
Y5
9
FIGURE 2.1
Eat at Your Own Risk
These are orange habaneros,
which rate about 300,000 on
the Scoville scale. The Scoville
scale is a psychophysical scale
that measures our experience
of piquancy or “hotness.” The
more capsaicin is present in a
particular food, the spicier it will
taste.
ISLE 2.1
North Carolina Hot
Sauce Contest Link
Psychophysical scale: a
scale on which people rate
their psychological experiences
as a function of the level of a
physical stimulus

29 Chapter 2: Research Methodology
these direct questions, we hope to understand the way
the mind works to accomplish sensation and perception.
Thus, the psychophysical approach focuses on the relation
between physical properties (e.g., the amount of capsaicin
present) and perception (the experience of hotness).
However, the actual questions asked of participants may
vary depending on the research focus. Participants may be
asked to do a detection task (“Is there any hotness in the
taste?”), a comparison task (“Which is hotter, Stimulus A
or Stimulus B?”), or a magnitude task (“How hot is the
stimulus?”). Moreover, the scale might be a preference
scale (“Which level of hotness is preferable?”). Researchers
need to be clear about which methods they use, because
sometimes different methods may reveal different patterns.
As such, it is important to detail the research methods of
sensation and perception in this chapter, before we delve
into individual sensory systems in the remaining chapters.
THE MEASURES AND
METHODS OF PSYCHOPHYSICS
2.1
Explain the nature of psychophysical scales and how they measure
the relations between real-world stimuli and our perceptions of them.
When you go to an optometrist to get your vision tested or to an audiologist to get your
hearing tested, you essentially perform a series of psychophysical tests. At the optom-
etrist, you are asked questions such as “Can you distinguish the d’s from the b’s and
the o’s from the q’s?” If you can, great; if you cannot, you may need to start wearing
glasses (Figure 2.2). At the audiologist, you are asked questions such as “In which ear
do you hear the tone?” If you can’t identify whether the sound is coming from the left
or the right, it may be time to consider hearing aids. But for our purposes, it is the basic
methods of psychophysics that allow us to start studying the processes of perception.
Many of the methods go back to Gustav Fechner in the 19th century.
Method of Limits
In the method of limits, stimuli are presented on a graduated scale, and participants
must judge the stimuli along a certain property that goes up or down. For example,
a participant may be presented with an increasingly dimmer set of lights. The par-
ticipant is asked to tell the experimenter when the lights are no longer visible. The
researcher will then present the participant with lights so dim that they cannot be seen
and then present increasingly intense lights until the participant detects them (Figure
2.3). Similarly, one could present a series of tones, starting at a volume so soft that one
cannot hear them and gradually increasing the loudness until the participant can hear
the tones. We illustrate the method of limits with ISLE 2.2. In the first screen of this
experiment, you will find a window in which you can set up the values that will adjust
how your method of limits experiment will run. The first item asks you to determine
the number of levels to test. The number of levels refers to the number of intensity steps
in the independent variable that will be tested. In the method of limits, the researcher
FIGURE 2.2 Woman Undergoing
a Routine Eye Exam at an Optometrist’s Office
People should have these exams annually to determine the
health of their eyes. Many of the tests an optometrist runs are
similar to psychophysical tests.
©
iStockphoto.com
/M
achineH
eadz
Method of limits: stimuli are
presented in a graduated scale,
and participants must judge the
stimuli along a certain property
that goes up or down

30 Sensation and Perception
hopes to pick an extreme value that is readily detected and a level that is never detected
and then several levels between these.
The method of limits is often used to determine both absolute and difference
thresholds. An absolute threshold is the smallest amount of a stimulus necessary to
allow an observer to detect its presence. For example, the smallest amount of light
energy at any particular wavelength of light that we can detect is its absolute thresh-
old. Similarly, the least amount of sound that we can hear at any particular frequency
is its absolute threshold. The smallest amount of capsaicin that we can detect as
hotness is its absolute threshold. Indeed, a bell pepper has such a small amount of
capsaicin that it earns a zero on the Scoville scale. For a final example, the smallest
amount of salt that your taste buds can detect is also its absolute threshold.
The method of limits can also be used to determine the difference threshold
(the JND mentioned earlier), which is the smallest difference between two stimuli
that can be detected. Thus, one might hold two weights and attempt to determine
if their masses are the same or different. The smallest difference in weight that can
be detected is the difference threshold, equal to 1 JND. Similarly, an observer might
see two green lights and be asked if the lights are the same or not in terms of expe-
rienced greenness. The smallest difference in the wavelengths of the lights that can
be detected is the difference threshold, or 1 JND. Similarly, the smallest increase of
capsaicin that can be detected as a difference in piquancy is also a 1 JND difference.
Returning to the concept of absolute thresholds, it turns out that detecting abso-
lute thresholds is harder than simply finding the softest sound a person can hear or
the dimmest light a person can see. For example, if you are just leaving the firing
range and have been hearing loud percussions for the past hour, detecting a very soft
tone might be difficult. If you have been studying in the library for the past hour, that
same tone might be quite audible. Similarly, it may be easier to detect a dim light in
the dark of night than it is to detect the same light on a bright summer day. If you
have just eaten a large and satisfying meal, the taste of another piece of lasagna may
not be as satisfying as the first. That is, absolute thresholds are not so absolute—they
depend on many internal and external conditions. In fact, our sensory systems are
constantly adapting to local conditions. On one hand, this makes assessing absolute
thresholds difficult, but on the other hand, it is an adaptive feature of our sensory
systems, as it allows us to perceive under a wide range of conditions. On a bright
sunny day at the beach, we want our sensitivity to light to be less than when we are
FIGURE 2.3
The Method of Limits
The participant must decide if a
stimulus is present at a number
of different levels of intensity.
The stimulus is increased in even
trials and decreased in odd trials.
A Y indicates that the participant
detects the stimulus, whereas an
N indicates that the participant
does not detect the stimulus.
The estimate of the threshold
is considered to be the mean
crossover point.
Intensity Trial Number
10 Y Y Y Y Y
9 Y Y Y Y Y
8 Y Y Y Y Y
7 Y Y Y Y Y
6 Y Y N Y Y
5 Y N Y Y Y Y
4 N Y N N Y Y N
3 N N N N N
2 N N N N
1 N N N N
Crossover 4.5 3.5 5.5 4.5 6.5 4.5 3.5 3.5 4.5
Threshold = mean crossover = 4.5
ISLE 2.2
Method of Limits
Absolute threshold: the
smallest amount of a stimulus
necessary to allow an observer
to detect its presence
Difference threshold (JND):
the smallest difference between
two stimuli that can be reliably
detected

31 Chapter 2: Research Methodology
trying to find our way around a forest campground on
a dark, moonless night.
In assessing thresholds, we often need to estimate
the threshold and try to compensate for any sensory
adaptation that is occurring, such as whether we have
just eaten. Thus, to determine an absolute threshold
with the method of limits, a researcher must use both
an ascending series and a descending series. An ascend-
ing series (or an ascending staircase) is one in which
a stimulus gets increasingly larger along a physical
dimension. Thus, the intensity of light might increase,
the amplitude of sound might increase, or the amount of capsaicin in a taste capsule
might increase. By contrast, a descending series (or a descending staircase) is one in
which the stimulus gets increasingly smaller along a physical dimension. Thus, the
researcher starts with a clearly visible light and lowers the amount of light on each
successive trial (Figure 2.4).
Consider a light detection experiment. We want to determine
a person’s absolute threshold for a red light in an otherwise dark
environment. In the ascending method, we start off with a level of
red light that is known to be below the threshold. The participant
should respond that he or she does not see it. The experimenter then
gradually increases the intensity of the light until the person can
detect the light. In the descending method, we start off with a bright
level of red light that the person can obviously detect and then lower
the intensity of the light until the person can no longer see it. The
point at which people change from detecting to not detecting or
vice versa is known as the crossover point. Typically, the threshold
will be different when measured by the ascending method versus
by the descending method. In general, with ascending series, people
are likely to claim that they can detect a stimulus when in fact the
stimulus is below the threshold. With descending series, people are
likely to claim that they cannot detect a stimulus when it actually
is above the threshold. Researchers will typically average over sev-
eral descending and ascending series to get their best estimate of the
absolute threshold (see ISLE 2.2).
Consider some common absolute thresholds in the natural world. Think of look-
ing up at the stars on a clear night (Figure 2.5). Think of the faintest star you can
possibly see—this star may be at or around your visual threshold. And for the audi-
tory system, consider hearing the faintest drone of a conversation from an upstairs
dorm room. You cannot make out the content of what they are saying, but you can
just barely hear their voices. In the domain of taste, consider how much sugar you
must put in your iced tea before it has the slightest hint of sweet. And think of the
faintest hint of a distant odor—perhaps a faraway smell of coffee. Thus, in real life,
we do encounter at-threshold stimuli from time to time.
In Chapter 1, we mentioned the work of Ernst Heinrich Weber, who did some
of the earliest work on thresholds in the 19th century. One area he investigated was
threshold in touch along the surface of the skin. He was interested in the two-point
touch threshold, which is the minimum distance along the skin at which two touches
are perceived as two touches and not one. In this case, two needles can be brought
gently to touch a person’s skin close to one another. If a person feels only one touch,
then it is below the threshold, but if the person feels two touches, then they are above
the threshold. Our two-point touch thresholds vary across our skin. Our fingers can
Ascending series: a series
in which a stimulus gets
increasingly larger along a
physical dimension
Descending series: a series
in which a stimulus gets
increasingly smaller along a
physical dimension
Crossover point: the point at
which a person changes from
detecting to not detecting a
stimulus or vice versa
Two-point touch threshold:
the minimum distance at which
two touches are perceived as
two touches and not one
FIGURE 2.4 Absolute Threshold
Illustration of the detection of absolute thresholds through the
method of descending limits. Each light is dimmer than the one to
its left.
©
iStockphoto.com
/Baltskars
FIGURE 2.5 Stars on a Clear Night
Many stars are easy to see, but some may lie just at
our thresholds. When we look at these stars straight
on, we may miss them, but we see them again “out
of the corner of our eye,” or at our periphery, which
is more light sensitive.

32 Sensation and Perception
detect two touches even when the needles are extremely close to each other. However,
the skin of our backs requires greater distance to feel two touches. Areas on the face
and mouth have small two-point thresholds, but not as good as the fingers. Other
areas, such as the arms and legs, require larger distances to perceive two touches, but
not as great distances as the skin of the back. The two-point threshold is an absolute
threshold. The size of the absolute threshold differs on different parts of the body,
just as acuity changes across the surface of the eye’s retina. Table 2.1 lists a few
everyday absolute thresholds.
TEST YOUR KNOWLEDGE
1. What is the difference between an absolute threshold and a difference
threshold? When might you want to use one and when might you want to use
the other?
2. Why would the method of limits not be a good method to test piquancy of hot
peppers?
Method of Constant Stimuli
In the method of constant stimuli, the threshold is determined by presenting the
observer with a set of stimuli, some of which are above the threshold and some of
which are below it; the stimuli are presented in a random order. This differs from the
method of limits. In the method of limits, the stimuli are changed to focus on a par-
ticular observer’s threshold. In the method of constant stimuli, all stimuli are always
presented, and all are selected beforehand. In the method of constant stimuli, the pre-
sentation of stimuli is given in a random order rather than zeroing in on the threshold.
This technique prevents the observer from being able to predict or anticipate what the
next stimulus will be. This reduces errors that may result from habituation or fluctua-
tions in perception due to attention or other factors. However, the method of constant
stimuli is often quite time-consuming, as it requires pretesting to gauge the general
region of the threshold for each participant. Moreover, it requires many trials to detect
a statistically determined threshold. In the method of constant stimuli, the stimulus that
is detected 50% of the time and not detected 50% of the time is considered to be the
threshold (Figure 2.6). You can see a demonstration of the method of constant stimuli
on ISLE 2.3.
The method of constant stimuli is used by audiologists when testing patients for
hearing thresholds (Gelfand, 2009). For each frequency of sound, the audiologist will
present the patient with an assortment of louder and softer tones to detect (see ISLE
2.3b). In this way, the audiologist can determine the threshold of hearing at each fre-
quency and, in some cases, if that threshold is higher than it should be, whether the
Method of constant stimuli:
a method whereby the threshold
is determined by presenting the
observer with a set of stimuli,
some above threshold and some
below it, in a random order
Sense Threshold
Vision A candle 30 miles away on a dark night
Audition A ticking watch 20 feet away in an otherwise silent location
Taste A teaspoon of sugar in 2 gallons of water
Smell A drop of perfume in three rooms
TABLE 2.1 Everyday Absolute Thresholds
Source: Adapted from Galanter (1962).
ISLE 2.3
Method of Constant Stimuli

33 Chapter 2: Research Methodology
person might be a candidate for hearing aids. Typically, hearing loss varies as a func-
tion of sound frequency, meaning that for some sound frequencies, a person may be
impaired but for others have normal hearing. An audiologist needs to know the
profile of hearing loss to properly program a patient’s hearing aids. For this reason,
as we will see later in the chapter, audiology starts with psychophysics.
Method of Adjustment
In the method of adjustment, the observer controls the level of the stimulus and
“adjusts” it to be at the perceptual threshold. The participant does so by increasing
or decreasing the level of the stimulus until it feels as if it is just at the detectable
level. Typically, the participant will do so by continuously adjusting a knob to control
the level of the stimulus. This is an intuitive measure for most participants because it
mirrors many normal activities, such as adjusting the volume control on a radio or a
dimmer switch on a light. In a light threshold study, the method of adjustment would
require the observer to adjust the light source to the dimmest light the participant can
detect. Adjusting the knob any lower would render the stimulus invisible. The advan-
tage of this technique is that it can quickly yield a threshold for each participant, but
a disadvantage is that it leads to great variance from one participant to the next and
between successive trials for each participant.
The method of adjustment is very useful for matching one stimulus to another to
determine the point of subjective equality (PSE). The PSE designates the settings of
two stimuli at which the observer experiences them as identical. For example, con-
sider a researcher who is interested in the JND in the detection of pitch differences
as a function of sound frequency. The observer hears a constant sample tone at, for
example, 1,000 Hz. The observer would then adjust a second tone until it matched
perceptually the sample tone. The experimenter could then look at the frequency
selected by the observer and see just how closely he or she actually matched the
sample tone. In vision research, the experimenter may present a stimulus of particu-
lar brightness. The participant would have to adjust another stimulus to be equally
bright to the first one. And, returning to our chili pepper example, an experimenter
might present a chili pepper sauce as the sample taste. Then a participant would have
to adjust the level of capsaicin in a second sample to match the level of hotness in
the first. To see and experience an auditory example of the method of adjustment, see
ISLE 2.4 and try a PSE in ISLE 2.5.
Stimulus level
100
75
50
25
0
5 10
Threshold
P
e
rc
e
n
ta
g
e
r
e
p
o
rt
e
d

S
e
e
n

15 20
100
75
50
25
0
5 10
I see it
I don’t see it
Stimulus level
P
e
rc
e
n
ta
g
e
r
e
p
o
rt
e
d

S
e
e
n

15 20
(a) (b) FIGURE 2.6
Measuring Threshold
These graphs illustrate how
we measure the threshold in
psychophysical experiments. In
Figure 2.6a, we see a hypothetical
cutoff at a particular level of
intensity that separates the stimulus
level at which we see the stimulus
and the stimulus level at which
we do not see the stimulus. Figure
2.6b illustrates that in most cases,
thresholds vary from trial to trial,
and we must estimate the threshold
from the point at which participants
are 50% likely to say “saw it” and
50% likely to say “didn’t see it.”
Method of adjustment: a
method whereby the observer
controls the level of the stimulus
and “adjusts” it to be at the
perceptual threshold
Point of subjective equality
(PSE): the settings of two
stimuli at which the observer
experiences them as identical
ISLE 2.4
Method of Adjustment
ISLE 2.5
Point of Subjective Equality

34 Sensation and Perception
Closely related to the concept of threshold is the concept of sensitivity. Sensitivity
is the ability to perceive a particular stimulus. It is inversely related to threshold.
As the threshold goes down, the observer is deemed to be more sensitive. That is,
lower thresholds mean higher sensitivity. This makes sense when one considers that
threshold refers to the ability to perceive a stimulus at smaller and smaller levels
of that quantity. Thus, a person who can hear a sound at 10 decibels (dB) is more
sensitive than a person who can hear the sound only at 15 dB. A person who can
smell a perfume at smaller concentrations in the air is more sensitive to that odor
than someone who requires larger concentrations. Sensitivity may vary across situa-
tions. For example, one may be more sensitive to changes in the intensity of light in
a room when one has just been in a dark room than when one has just come out of
the brilliant sunshine.
Magnitude Estimation
Magnitude estimation is a psychophysical method in which participants judge and
assign numerical estimates to the perceived strength of a stimulus. This technique was
developed by S. S. Stevens in the 1950s (e.g., Stevens, 1956). Magnitude estimation usu-
ally works in the following way. An experimenter presents a standard tone and assigns
it a particular loudness value, say 20. Then the participant must judge subsequent tones
and give them numerical values, in comparison with the standard. So if the participant
thinks the new tone is twice as loud as the standard, it should be assigned a 40. If the
next tone is just a bit softer than the standard, it may be assigned a 15. If the tone is
heard to be much softer than the standard, it might receive a 5 on this hypothetical
scale of loudness.
Magnitude estimation can also be used for visual experiments. It follows the same
principle: In brightness estimation, a sample light might be given a standard value,
and then other samples are judged in relation to the standard. Magnitude estimation
can be adapted to just about any perceptual dimension. Participants may be asked to
judge the brightness of light, the lengths of lines, or the sizes of circles on a numerical
scale. For visual and auditory examples of magnitude adjustment, go to ISLE 2.6.
As experimenters in a psychophysics experiment, we can control the stimulus. For
example, we can have people taste and make judgments of the sweetness of a sample
with a known quantity of sugar, say 1 teaspoon per gallon. We then present them
with a sample of a solution of sugar water calibrated to 2 teaspoons per gallon. The
new stimulus has twice as much sugar in it per unit of water, but is it perceived as
twice as sweet? In all sensory systems, there is a phenomenon called response com-
pression. Response compression means that as the strength of a stimulus increases,
so, too, does the perceptual response, but the perceptual response does not increase
by as much as the strength of the stimulus increases. That is, if you judge the first
sugar-water solution as a 5 on the sweetness scale, doubling the sugar in the solution
will cause an increase in your sweetness judgment but not a doubling of the perceived
sweetness. In fact, you will likely judge the 2-teaspoon solution as a 7 or an 8 on the
sweetness scale.
There is an exception to the response compression rule for sensory perception,
and that involves pain perception. In pain perception, there is response expansion
instead of response compression. Response expansion means that as the strength
of a stimulus increases, the perceptual response increases even more. In this case,
if a person receives an electric shock of physical intensity 10 and judges it to be
5 on a pain scale, increasing the physical intensity to 20 will lead to a more than
doubling of the judgment, to perhaps 15 on the pain scale. That is, smaller incre-
ments of increase in the physical dimension (e.g., electric shock) lead to greater
ISLE 2.6
Magnitude Estimation
Sensitivity: the ability to
perceive a particular stimulus; it
is inversely related to threshold
Magnitude estimation: a
psychophysical method in which
participants judge and assign
numerical estimates to the
perceived strength of a stimulus
Response compression:
as the strength of a stimulus
increases, so does the
perceptual response, but the
perceptual response does not
increase by as much as the
stimulus increases
Response expansion: as
the strength of a stimulus
increases, the perceptual
response increases even more
Stevens’s power law: a
mathematical formula that
describes the relationship
between stimulus intensity and
our perception; it allows for
both response compression
and response expansion

35 Chapter 2: Research Methodology
increments in the perception of the perceptual characteristic (e.g.,
the perception of pain).
Stevens (1957, 1961) developed an equation to try to encapsu-
late both types of data sets (this is one of just a few mathematical
equations that will be presented in this book). It is called Stevens’s
power law, and it is as follows:
P = cIb
In this equation, P is equal to the perceived magnitude of a stim-
ulus, that is, how bright we perceive a light to be or how sweet we
perceive a sugar solution to be. I is equal to the intensity of the actual
stimulus. Thus, at the simplest level, our perception is a function of
the physical intensity of the stimulus. However, there are two other
parts of the equation that explain the relation between perception
and the physical stimulus. The letter c represents a constant, which
will be different for each sensory modality or for each sensory dimen-
sion. The constant also allows you to scale your measure appropri-
ately. For example, both Fahrenheit and Celsius temperature scales
measure the same underlying property, but they do so with different
scales. The exponent b equals the power to which the intensity is
raised. It is this exponent b that allows response compression and
response expansion. Response compression occurs when b is less
than 1, and response expansion occurs when b is greater than 1.
Thus, Stevens’s power law equation can account for both types of
subjective responses. This is depicted in Figure 2.7. For a graphical
demonstration of how this works, go to ISLE 2.7.
To give an example, let us return to the Scoville scale. We could
present a chili sauce with 1 milligram of capsaicin per kilogram
of other substances. We tell people that this is a 1 on the piquancy scale. We then
increase the dosage to 2 mg/kg. Thus, we have doubled the active “hot” ingredient.
Applying Stevens’s law, we need to know the constant and the exponent to determine
a person’s perceptual response. Because capsaicin stimulates pain receptors, we can
expect an exponent of greater than 1 (and hence response expan-
sion). Thus, the hotness judgment should be more than double
the baseline, leading to a hotness judgment of greater than 2. We
must know the psychological response in order to apply Stevens’s
law. That is, knowing the amount of capsaicin in a food would
not have allowed Scoville to measure piquancy. He needed the psy-
chophysical measure to map response expansion. A few exponents
for Stevens’s power law are given for different sensory domains
in Table 2.2. A value greater than 1 indicates response expansion,
whereas a value less than 1 indicates response compression.
Catch Trials and Their Use
One of the potential difficulties that arise with the traditional methods
of limits, constant stimuli, and adjustment is that the participant might
be willingly or unwillingly misinforming the experimenter about perceptual experience.
For example, a participant may indicate that he heard a sound when he was not sure
or perhaps because he thinks a sound should have occurred even though he did not
perceive one. Or worse, perhaps the participant wants to impress the experimenter with
100
75
50
25
0
Electric shock
Bar length
Brightness
Physical intensity of
stimulus
P
e
rc
e
iv
e
d
i
n
te
n
si
ty
o
f
st
im
u
lu
s
FIGURE 2.7 Comparison of the Physical
Intensity of a Stimulus and Its Perceptual
Correlate
In response compression, as the physical intensity
of a stimulus increases, its perceptual correlate
increases as well, but not by as much. In response
expansion, as the physical intensity of a stimulus
increases, its perceptual correlate increases even
more. The curve for brightness illustrates response
compression, whereas the curve for electric shock
illustrates response expansion
ISLE 2.7
Stevens’s Power Law
Sense Perception Exponent
Vision Brightness 0.3
Audition Loudness 0.5
Taste Sweetness 0.8
Vision Apparent length 1.0
Touch Pain 3.5
TABLE 2.2
Stevens’s Power Law Exponents
Source: Adapted from Stevens (1961).

36 Sensation and Perception
his extraordinary sensory abilities. In all of the methods described so
far, it would be easy to do; you could simply report that you see or hear
the stimulus when you do not. Because the stimulus is always present,
even at a very soft volume or at a very low brightness, the experimenter
cannot know if the participant is being truthful or not. One technique
to counter this strategy is to use catch trials. A catch trial is a trial in
which the stimulus is not presented. It is easy to insert these trials as
checks on the participant’s accuracy and honesty. Thus, in catch trials,
the correct answer is “No, I didn’t hear it” or “No, I didn’t see it.” If a
participant reliably says that she saw the stimulus in a catch trial, we
can dismiss this participant as an unreliable observer.
Another method that circumvents problems of false reporting is
the forced-choice method (Blackwell, 1953; Jones, 1956). In every
trial, the subject is asked to report either when the stimulus occurred
or where it occurred. Thus, instead of determining whether a light
was present or not, the participant must decide if a light was present
in one location or another or at one time slot or another time slot.
Instead of determining whether the participant heard a sound or not,
the participant must determine in which of two time intervals there
was a sound or not. This technique prevents the need for catch trials
because the observer cannot simply say “yes” in every trial, regard-
less of the presence of a stimulus in that trial. But it also allows the
determination of thresholds, because if a stimulus cannot be detected, performance
will be at chance (50% if there are two choices). Threshold can be determined by
finding a level of performance that is significantly above chance. See ISLE 2.8 for an
illustration of the forced-choice method (also see Figure 2.8).
TEST YOUR KNOWLEDGE
1. What is the point of subjective equality (PSE)? Why is it critical when using the
method of adjustment?
2. How can the equation to represent Stevens’s power law explain both response
compression and response expansion?
2.2 Demonstrate an understanding of signal detection theory.
Signal Detection Theory
On July 3, 1988, Iran Air Flight 655 was shot down by a U.S. missile from the
Navy vessel U.S.S. Vincennes (Figure 2.9). All 290 passengers and crew, including
66 children, were killed. Although the United States has never apologized to Iran,
the U.S. government gave $61 million to the relatives of the victims of the attack.
The incident took place at a time when there was heavy tension between the United
States and Iran, and Iranian jets had previously attacked U.S. Navy vessels. However,
in this case, the radar operators on the Vincennes mistakenly judged the civilian
airplane, an Airbus 300, to be an incoming Iranian F-14 fighter, with the horrify-
ing consequences described here. How had the Navy made such a terrible mistake?
We will couch an explanation of how this disaster occurred in terms of one of the
most influential theories in the history of sensation and perception research, namely,
signal detection theory.
Consider the radar specialist examining the screen that depicted incoming objects.
He or she had to decide, on the basis of the information available on the radar screen,
ISLE 2.8
Forced-Choice Method
Catch trial: a trial in which the
stimulus is not presented
Forced-choice method: a
psychophysical method in which
a participant is required to
report when or where a stimulus
occurs instead of whether it
was perceived
Signal detection theory: the
theory that in every sensory
detection or discrimination,
there is both sensory sensitivity
to the stimulus and a criterion
used to make a cognitive
decision
FIGURE 2.8 Illustration of the
Forced-Choice Method
The participant in this experiment is engaged in
a psychophysical task using the forced-choice
method, pressing the right or left button to
indicate where a light displays on the screen.

37 Chapter 2: Research Methodology
whether the incoming object was an enemy warplane (an
Iranian F-14) or a harmless jetliner (an Airbus 300). In
1988, this involved examining a radar screen and making
a judgment as to the nature of the incoming airplane. The
United States had (and has) complicated procedures to dif-
ferentiate military and civilian airplanes on incoming radar.
But in a war zone with an object approaching you at 600
mph, decisions must be made quickly.
From the U.S. Navy’s standpoint, there are two types
of errors. The Navy could mistake a civilian airplane for
a military jet, or it could mistake a military jet for a civil-
ian airplane. Both of these errors could have fatal conse-
quences for innocent people, as indeed they did in this case.
In the Vincennes case, the mistake that arose is called a
false alarm. A false alarm, in psychophysical terms, occurs
when the observer mistakes a harmless or null signal for
a dangerous or an active signal. The other type of error
is called a miss, in which a harmful signal is perceived as
harmless. In one case, the danger is to innocent civilians aboard the aircraft, whereas
in the other case, the danger is to equally innocent personnel aboard the Navy vessel.
Thus, in this situation, everything must be done to avoid both errors (Table 2.3).
In such a situation, there are also two potential correct responses. A correct rejection
occurs when a harmless signal is perceived as harmless, and a hit occurs when a
harmful signal is correctly perceived as harmful. In the case of the Vincennes, the
correct response should have been a correct rejection. The plane was a civilian jet-
liner that offered no threat to the Navy. In this example, a hit would have occurred if
the radar had correctly identified an incoming military warplane as a threat. Military
action must optimize hits and correct rejections in order to achieve strategic objec-
tives and minimize civilian casualties. Thus, one of the goals of any military surveil-
lance equipment is to maximize hits and correct rejections while minimizing false
alarms and misses.
Consider a radar operator examining the signal on his or her radar screen. Such
an operator may be more likely to define a signal as “dangerous” if the signal is
coming from an aircraft over the Strait of Hormuz, especially after recent attacks on
the U.S. Navy, than if he or she is monitoring aircraft flying over the skies of central
Nebraska. This differing judgment on the basis of situation is called the criterion. In
Nebraska, the radar specialist will select a very high criterion, so as to avoid false
alarms, and misses are less of a concern. However, when a Navy vessel is patrolling
the Strait of Hormuz, near a highly hostile Iranian government, which has recently
launched attacks against U.S. ships, the criterion will be lower, as there are more
risks in the environment and the Navy does not want to miss an actual attack; thus,
False alarm: in signal detection
analysis, a false alarm is an
error that occurs when a
nonsignal is mistaken for a
target signal
Miss: in signal detection
analysis, a miss is an error that
occurs when an incoming signal
is not detected
Correct rejection: in signal
detection analysis, a correct
rejection occurs when a
nonsignal is dismissed as not
present
Hit: in signal detection analysis,
a hit occurs when a signal is
detected when the signal is
present
Criterion: a bias that can
affect the rate of hits and false
alarms
Incoming Signal
Response Is a Fighter Jet Is a Civilian Plane
Think it is a fighter jet Shoot down enemy plane Innocent civilians killed
Think it is a civilian plane Innocent sailors killed No hostile interactions
TABLE 2.3 Signal Detection Theory: Is It a Threat?
FIGURE 2.9 The U.S.S. Vincennes
The U.S.S. Vincennes, seen here in 2005, was a guided-missile
cruiser. In 1988, its crew mistakenly shot down a civilian
aircraft. This error, called a false alarm in psychophysical
terminology, had devastating consequences.
A
P Photo/G
reg English
ISLE 2.9
Signal Detection Experiment

38 Sensation and Perception
there is greater risk for more false alarms. This is exactly what happened off the
coast of Iran in 1988. Thus, this military tragedy fits into the logic of signal detection
theory, as we will shortly see (Green & Swets, 1966). See ISLE 2.9 to play such a
war game yourself.
Here’s another way of thinking about signal detection theory (Figure 2.10).
Consider the following situation: You are driving down the road, and you think
you hear a noise from the engine that sounds like a clunk. However, with your
stereo playing and the sounds of the road, you are not sure exactly what the clunk
is. If your car is relatively new and has a history of smooth running, perhaps you
will decide that you didn’t really hear anything (high criterion). Your hearing is
“playing tricks” on you. However, if you are driving an old car that has a history
of spending nearly as much time in the shop as on the road, you might decide
that you did hear something and head to the nearest service center (low criterion).
What is important in this example is that even in this very basic sensory discrim-
ination, there is a cognitive decision-making element that needs to be taken into
account. This is another practical example of the situations for which signal detec-
tion accounts.
So let us look at a model of signal detection starting with a visual detection
example (Table 2.4). Signal detection assumes that there is “noise” in any system.
On occasion, this noise may be mistaken for an actual signal. In these examples,
what should have been seen as noise, in the military sense, was mistaken for a mili-
tary signal by the crew of the Vincennes. This noise may occur because of distortion
on the screen or a wobble on the radar caused by atmospheric conditions. Similarly, if
we have an old car, we may hear
“clunks” even when the car is
operating effectively. This noise
in the system may be other sounds
from the road, tinnitus in our
ear, or something rustling in
the trunk. Thus, the task for
the observer is to distinguish
an actual signal from the back-
ground noise. In a vision thresh-
old experiment, this might be
the actual dim light in the noise
produced by random firing of
retinal receptors.
So consider first the case in
which there is a faint signal. We
can detect it (hit) or not detect it
(miss). But there is also the case
in which the signal is absent. In
that case, we may detect some-
thing (false alarm) or correctly
realize that there is no signal
(correct rejection). Signal detec-
tion theory states that we must
consider all four possibilities to
successfully measure thresholds.
FIGURE 2.10
Old Car/New Car
When we think we hear a noise
associated with an engine
malfunction, we adopt different
criteria, depending on the situation.
If we are driving the old clunker
(a), we may adopt a more liberal
criterion for hearing an engine
problem. This will help us catch
engine trouble (hits) but may also
increase the rate of bringing the
car to the shop when there is no
problem (false alarm). If we are
driving the shiny new sports car (b),
we may adopt a more conservative
criterion for hearing an engine
problem. This will increase the
likelihood of correctly dismissing
a sound as noise rather than a
mechanical problem (correct
rejection) but may increase the
likelihood that we do not hear a
problem when there is one (miss).
©
iS
to
ck
Ph
ot
o.
co
m
/n
w
in
te
r
©
iS
to
ck
Ph
ot
o.
co
m
/R
aw
pi
xe
l
(a)
(b)

39 Chapter 2: Research Methodology
Consider a situation in which there is little noise in the system. Imagine if the
captain of the Vincennes in 1988 had access to 2018 satellite imaging (perhaps cour-
tesy of Dr. Who, the fictional time-traveling British superhero). In this case, there
is little noise in the system. He could watch the video feed from the satellite and
see that the incoming plane is an Airbus rather than an F-14. In this case, there
would be few false alarms and few misses, only hits and correct rejections. But in
1988, the radar technology had noise in the system, presumably generated by any
number of atmospheric conditions. Thus, the noise introduces the potential for error
in the system.
These differences are represented graphically in Figure 2.11. In the figure, we con-
sider a common psychophysical task, detecting a soft sound in a quiet environment.
The Signal Is
Response Present Absent
Perceived it Hit False alarm
Did not perceive it Miss Correct rejection
The Clunk
Response Actual Did Not Happen
Think it happened You get needed service You make an unneeded
service visit
Think it did not happen You break down You go happily on your way
FIGURE 2.11
Signal Detection Theory
Illustrated Graphically
Figure 2.11a shows the
continuum of neurons firing
in response to a soft or an
absent sound. Figure 2.11b
shows a potential distribution
for perceiving the soft sound or
misperceiving the absent sound.
Figure 2.11c adds the criterion.
Figures 2.11d through 2.11g show
correct rejections, hits, false
alarms, and misses.
N
u
m
b
e
r
o
f
in
st
a
n
ce
s
N
u
m
b
e
r
o
f
in
st
a
n
ce
s
Less More
Your perception
Less More
Sounds like phone
Less More
Sounds like phone
NO YESCriterion
NO YESCriterionNO YESCriterionNO YESCriterionNO YESCriterion
(d) Correct rejection (e) Hit (f) False alarm (g)
(a) (b) (c)
Miss
Less More
Sounds like phone
Less More
Sounds like phone
Less More
Sounds like phone
Less More
Sounds like phone
Shower “noise” alone
Ring + noise
TABLE 2.4A Signal Detection Theory: Possible Situations
TABLE 2.4B Signal Detection Theory: Possible Situations

40 Sensation and Perception
Noise here can be defined as random neuronal firing—think of the odd “sound” you
experience but are not sure if it is for real. Thus, a hit occurs when we detect a soft
sound when present, and a false alarm represents when we think we heard a soft
sound when one was not present. A correct rejection occurs when we say that no
sound was present when no sound was present, but a miss occurs when we say that
no sound was present when in fact there was a soft sound. Each of these situations
is represented in the figure.
Let’s return now to an idea introduced in comparing the threat in the Strait of
Hormuz and the threat in Nebraska, that is, the idea of a criterion in signal detec-
tion theory. The criterion is an internal cutoff determined by the observer, above
which the observer makes one response and below which the observer makes another
response. Translated into technical terms, then, the criterion is a bias that can affect
the rate of hits and false alarms. Thus, in the Strait of Hormuz, we adopt a criterion
or an internal cutoff to respond at a lower threat level than we do in Nebraska. After
all, we assume that the danger of an incoming enemy aircraft is much higher off the
coast of a hostile nation than in the safety of the middle of our own. In our other
example, if the clunking sound is above a certain level of loudness, we decide that
there really is a clunk, and there is a problem with the car. If it is below that level
of loudness, we decide that there is no reason to worry. This cutoff point may be
different depending on whether we have a new car or that old clunker. Similarly, in
a psychophysical study, the criterion determines the level of stimulation above which
we decide a light is present and below which we decide not to indicate the presence
of the light.
However, criteria vary depending on the situation. You might adopt a lax crite-
rion for hearing the clunk if your car is old and has a history of engine problems.
That is, you are more likely to attribute the clunk to the car than to random road
sounds. This means you will risk more false alarms to catch all hits—you don’t want
your car to fall apart at 70 mph on the highway. But there is little cost to taking it
the shop one more time. Alternatively, if your car is new and has just been checked
by a mechanic, you may adopt a stricter strategy. This means that you are more likely
to attribute the noise to the road and not to the car. In signal detection terms, this
means a greater desire to avoid false alarms (not
paying extra money to the mechanic) and risking
missing hits (i.e., a breakdown, which you think
is unlikely).
Thus, a criterion is a value set by the observer
depending on circumstances that only he or she
knows. If the stimulus is above the criterion, the
observer will say “yes” or “present,” but if the
stimulus is below the criterion, the observer will
say “no” or “absent.” Consider how the concept
of a criterion might apply to real-world decisions.
A radiologist screening for breast cancer must
weigh the risks of unnecessary testing against the
risk for cancer in evaluating a mammogram. If
the woman is young and has no family history
of breast cancer, the radiologist might adopt a
stricter criterion, not wanting to risk a false alarm
and put the woman through the pain of taking a
FIGURE 2.12 Sensitivity
Sensitivity is the ability to distinguish a noise signal from an actual signal,
such as the difference between random noise and the clunking sound in
your car. Figure 2.12a shows an example of when the perceiver has no
sensitivity—it is impossible to tell the difference between signal and noise.
Figure 2.12b shows when the perceiver is somewhat better at distinguishing
signal from noise, and Figure 2.12c shows much higher sensitivity.
N
u
m
b
e
r
o
f
in
st
a
n
ce
s
N
u
m
b
e
r
o
f
in
st
a
n
ce
s
N
u
m
b
e
r
o
f
in
st
a
n
ce
s
Less More Less More
Sounds like phone
d’ = ∼4d’ = ∼1d’ = ∼0
Sounds like phoneSounds like phone
Less More
(a) (b) (c)No sensitivity Moderate sensitivity High sensitivity
Shower “noise” alone
Ring + noise
Sensitivity (signal detection
theory): the ease or difficulty
with which an observer can
distinguish signal from noise
d′ (d-prime): a mathematical
measure of sensitivity
Receiver-operating
characteristic (ROC) curve: in
signal detection theory, a plot of
false alarms versus hits for any
given sensitivity, indicating all
possible outcomes for a given
sensitivity

41 Chapter 2: Research Methodology
tissue sample. However, if the woman is older or
has a family history of breast cancer, the radiolo-
gist may adopt a more lax criterion, not wanting
to make a miss, and recommend further testing.
Mammographic screening is very effective, but
there are still misses and false alarms to be con-
cerned about.
There is one last piece to understand in sig-
nal detection theory. Sensitivity, as used in sig-
nal detection theory, is the ease or difficulty with
which the observer can distinguish the signal
from noise. That is, sensitivity measures how
easy it is to tell if a signal is present or absent
(Figure 2.12). Thus, one can imagine a radar
operator with 1948 technology and one with
2018 technology. The older system has lower sen-
sitivity, so it is harder to determine the nature of
the incoming threat. The 2018 version has higher
sensitivity, so it is easier to tell the difference between a harmless
passenger plane and a dangerous enemy fighter jet. Similarly, if
you have had a “clunking” problem before, you may know exactly
what to listen for. Therefore, it will be easy to distinguish between
noise and the clunk.
An observer with high sensitivity will be able to make mostly
hits and correct rejections. But, as you can see in Figure 2.12,
as sensitivity decreases, more false alarms and misses occur.
As sensitivity increases, the observer has more hits and correct
rejections. Thus, when you know what distinguishes between
noise and clunks, you will make few mistakes and catch the
car if it is on the brink of breaking down and ignore the
sound if you know that it is just random noise. In psychophysics,
if you know the relation of hits to false alarms, you can
determine d′ (d-prime), which is a mathematical measure of
sensitivity.
Sensitivity and criterion interact in interesting ways.
Sensitivity may be high, but if the criterion is relatively low,
there still might be many false alarms. Even if sensitivity is
high, if the criterion is very loose, there still might be many
false alarms. That is, even if our radar system is very good at
detecting enemy aircraft, a trigger-happy officer might still be
making too many false alarms. This is illustrated in Figure 2.13.
You can also review all of the issues related to signal detection
theory in ISLE 2.10.
For any given sensitivity d′, there is a range of possible out-
comes according to signal detection theory. To simplify seeing all of the possible out-
comes for a given signal strength, researchers have developed a way to summarize all
of the possible outcomes for this situation across all possible criteria. This summary
is called the receiver-operating characteristic (ROC) curve (Figure 2.14). The ROC
curve is a graphical plot of how often false alarms occur versus how often hits occur
FIGURE 2.13 Hits and False Alarms
When sensitivity is kept constant, there can still be differences in the ratio
of hits to false alarms, depending on the criterion. Figure 2.13a shows a lax
criterion, which allows many false alarms but maximizes hits. Figure 2.13b
shows a medium criterion, and Figure 2.13c shows a strict criterion, which
minimizes false alarms but also reduces the detection of hits.
Shower “noise” alone
Ring + noise
N
u
m
b
e
r
o
f
in
st
a
n
ce
s
N
u
m
b
e
r
o
f
in
st
a
n
ce
s
N
u
m
b
e
r
o
f
in
st
a
n
ce
s
Less More Less More
Sounds like phoneSounds like phoneSounds like phone
Less More
(a) (b) (c)“Gotta get that call!” “Is that the phone?” “What phone?”
FIGURE 2.14 Receiver-Operating
Characteristic (ROC) Curves for Different Values
of d′ (Sensitivity)
When d′ = 0, sensitivity is zero, and the perceiver
cannot discriminate between signal and noise. As d′
increases, hits and correct rejections increase, and
misses and false alarms decrease. If d′ were perfect,
we would get only hits and correct rejections, but in
the real world, d′ is never perfect.
Correct rejection Pr (N/n)
False alarm Pr (S/n)
1.0
1.00
0
1.0 0
01.0
H
it
P
r
(S
/s
)
M
iss P
r (N
/s)
d’ = 2
d’ = 1.5
d’ = 0.75
d’ = 0
Ch
an
ce
d
ia
go
na
l
ISLE 2.10
Signal Detection Theory

42 Sensation and Perception
for any level of sensitivity. See ISLE 2.11, in which you can adjust the criterion and
the sensitivity and see what the ROC curve looks like.
The advantage of ROC curves is that they capture all aspects of signal detec-
tion theory in one graph. Sensitivity, or d′, is captured by the “bow” in the curve.
The more the curve bends up to the right, the better the sensitivity. Moving
along the bow captures the criterion. As one moves up any individual level
of sensitivity, the higher you go, the more loosely you are setting your criterion,
thus opening the possibility of more hits but also more false alarms. Regarding
Figure 2.14, the blue line indicates the highest d′. This observer can set her crite-
rion low and still catch many hits while minimizing false alarms. For the observer
represented by the green line to make as many hits, he risks more false alarms.
If the blue line and green line were radar operators, we would prefer to have Blue
line making decisions, as she has a higher sensitivity. Depending on circumstances,
she can set her criterion low or high and achieve the accuracies indicated on the
ROC curve.
ROC curves are also important in evaluating medical decision making.
Consider radiologists evaluating computerized tomographic (CT) scans looking for
brain tumors. They want to be able to detect as many brain tumors as possible but
also minimize the risk for false alarms, as false alarms can be quite dangerous—one
does not want to initiate brain surgery if there is
nothing wrong with the patient’s brain. So in eval-
uating radiologists, we want those with the highest
values of d′, that is, those who can best distinguish
real tumors from false alarms in the CT scans. And
we may also want them to set intermediate crite-
ria, because false alarms also come with risk (Xue,
Peng, Yang, Ding, & Cheng, 2013).
TEST YOUR KNOWLEDGE
1. What is signal detection theory? How does it
differentiate between accurate perception and
errors?
2. What is d′ in signal detection theory? Why is it
important in graphing ROC curves?
2.3
Evaluate neuroscience methods and
what they tell us about
sensation and perception.
Neuroimaging Techniques
Neuroimaging techniques are technologies that permit visual examination of living
human brains. These techniques allow scientists to correlate perception with brain
activity. Neuroimaging techniques are also used to investigate memory, attention,
problem solving, and emotion. Two goals of neuroimaging are to reveal where per-
ception happens in the brain and how perception unfolds through the brain over
time. To reveal the particular area(s) in the brain where a specific process occurs,
scientists use these methods to develop spatial maps of the brain, which show the
FIGURE 2.16 Magnetoencephalography (MEG)
In MEG, magnetic sensors detect magnetic fields produced by the
electrical activity in the brain. MEG has better spatial resolution than EEG.
Th
e
Sy
dn
ey
M
or
ni
ng
H
er
al
d/
Fa
irf
ax
M
ed
ia
/G
et
ty
Im
ag
es
FIGURE 2.15
Electroencephalography (EEG)
In EEG, electrodes monitor the
electrical output of large numbers
of neurons in the brain. The more
electrodes that are placed on the
scalp, the better the ability will
be of the EEG to specify spatial
locations in the brain.
©
iS
to
ck
ph
ot
o.
co
m
/f
ot
og
ra
fix
x
Electroencephalography
(EEG): using electrodes to
measure the electrical output of
the brain by recording electric
current at the scalp
ISLE 2.11
Signal Detection Theory
and the Receiver-Operating
Characteristic (ROC) Curve

43 Chapter 2: Research Methodology
areas that are active during perceptual tasks. To ascertain
changes in activity in the brain over time, scientists use
neuroimaging pictures in quick succession to find the time
course of perceptual processes. The field of neuroimag-
ing is changing quickly, but we review four of the main
techniques.
1. Electroencephalography (EEG). In EEG, up to 128
electrodes are positioned on a person’s scalp (Figure
2.15). The electrodes detect electrical signals created
by the brain. Areas of the brain that are active gener-
ate faster changing electrical signals than those that
are not active, and thus, by comparing electrical out-
put across electrodes, we can approximate where a
perceptual process is coming from in the brain. EEG
picks up a continuous electric signal, which allows
for measurements to be made every millisecond.
Thus, EEG allows for determining the time course of
perceptual processes in the brain.
2. Magnetoencephalography (MEG). In MEG, magnetic sensors detect small mag-
netic fields produced by the electrical activity in the brain (Figure 2.16). MEG
detects rapid changes in the brain, although at a less precise time scale than
does EEG. However, MEG produces better spatial maps of the brain than EEG,
though not as good as fMRI, discussed next.
3. Magnetic resonance imaging (MRI) and functional magnetic resonance
imaging (fMRI). In this technique, large magnetic fields align the oxygen
molecules within our brains (Figure 2.17). Then, as blood flows into
areas of the brain, the molecules’ organization is disrupted, which can be
detected by sensors in the MRI machine. The fMRI scanners take a pic-
ture approximately every 30 milliseconds, which allows scientists to
pinpoint both where and over what time course perceptual processes
are happening in the brain. Such fMRI scans are incredibly important
right now in determining the areas of the brain responsible for various
perceptual processes and the connections in the brain that link different
areas involved in the complex process of turning sensory stimulation into
experiences.
4. Transmagnetic Stimulation (TMS). We include a number of technologies
that can be grouped together under the general label of TMS (Figure 2.18).
These techniques stimulate the brain via electric current, but the specifics
of the technique and what kind of current use varies from one technique to
another. In TMS, researchers place a magnetic field generator (or coil) on the
head of participants. The coil induces an electric current in the particular
brain region beneath the coil, which induces changes in how this region func-
tions, which may be either perceptual or behavioral. In research, these changes
are temporary, but they permit the experimental analyses of brain region and
function. Indeed, a TMS pulse to the occipital lobe can induce temporary
blindness.
Magnetoencephalography
(MEG): using a magnetic
sensor to detect the small
magnetic fields produced by
electrical activity in the brain
Transmagnetic stimulation
(TMS): a procedure in which
a magnetic coil is used to
stimulate electrically a specific
region of the brain
FIGURE 2.17 Magnetic Resonance Imaging (MRI)
Large magnetic fields align the oxygen molecules within
our brains. Then, as blood flows into areas of the brain, the
molecules’ organization is disrupted, which can be detected by
sensors in the MRI machine.
BSIP/U
niversal Im
ages G
roup/G
etty Im
ages
FIGURE 2.18
Transmagnetic Stimulation
(TMS)
TMS techniques stimulate
the brain via electric current.
Researchers can see if perception
or behavior changes in a particular
domain as a function of that
stimulation, allowing researchers
to draw causal relations between
that brain region and perception.
G
aro/Phanie/Science Source

44 Sensation and Perception
EXPLORATION: Intersensory Perception
The chapters in this textbook cover vision, audition, the skin
senses, and the chemical senses. In these chapters, we discuss
the underlying neuroanatomy, the physiology, and the psy-
chological science of each area. However, real people mov-
ing and acting in the world are processing all of these senses
simultaneously. In many domains, understanding what is
going on in the world involves co-perceiving across two
modalities. For example, when we talk with someone, we
both listen to their words and watch their faces. When we eat
our dinner, we simultaneously taste our food, smell our food,
look at our food, feel the texture of the food in our mouths,
and, as discussed at the beginning of the chapter, feel the pain
caused by certain chemicals in our food. A natural question
arises as to whether or not one sensory system affects other
sensory systems, and, if so, how? This Exploration section
looks at some research that uses psychophysical measures to
address how olfaction (sense of smell) affects visual thresh-
olds (Robinson, Laning, Reinhard, & Mattingley, 2016).
In order to work through the work examining the effects
of odor on visual thresholds, a short digression is needed
concerning the nature of visual masking. Masking refers
to the difficulty in seeing one stimulus when it is quickly
replaced by a second stimulus that occupies the same or
adjacent spatial locations. In masking, a briefly shown
and to-be-detected stimulus that would normally be visi-
ble is rendered invisible by the presence of a second stim-
ulus that occurs in the same location immediately after
it or by stimuli that surround the test stimuli and per-
sist after the test stimulus disappears (Enns & Di Lollo,
1997). Because masking is not easy to imagine, please
take a moment to log onto ISLE and view the demon-
stration provided there (ISLE 2.12).
In a masking experiment, a to-be-detected stimulus is pre-
sented to participants somewhere in their visual field. The
stimulus may be a simple dot, a photograph of a face,
or even a word. The
to-be-detected stimulus
is presented for a very
short time span, perhaps
50 milliseconds. This is
long enough that if the stimulus is presented by itself, most
people will report seeing it. However, if the stimulus is
immediately replaced by another stimulus in the same area,
the person will not consciously detect the to-be-detected
stimulus. This is a phenomenon known as backward
masking (Breitmeyer & Hanif, 2008). Masking can also
occur if the to-be-detected stimulus is surrounded by other
stimuli that are presented at the same time and persist a
bit longer than the to-be-detected stimulus. If the masking
stimuli disappear at the same time as the to-be-detected
stimulus, the to-be-detected stimulus will be detected.
However, if the masking stimuli persist after the to-be-
detected stimulus, then it will be masked and consciously
invisible. This form of masking is called object-substitu-
tion masking (Enns & Di Lollo, 1997).
Because masking experiments require participants to
make decisions about whether a visual target is present
or not, they lend themselves to signal detection analyses.
In signal detection analyses, how good a person is at
detecting the stimuli is a measure of sensitivity. Indeed,
studies using masking usually report participants’ sensi-
tivity in terms of d′ analyses, which we covered in this
chapter. A high d′ means that participants are good at
detecting the stimulus when it is present but also good
at rejecting a trial when the stimulus is not present. This
brings us back to the topic of this section: Does the
presence of stimuli from another sensory domain affect
visual sensitivity? We can answer this question by look-
ing at the sensitivity of participants in a masking exper-
iment when stimuli from another sensory domain are
present or not. This question means presenting stimuli
from another sensory domain such as audition or olfac-
tion before or during a masking trial and determining if
the presence of the second sensory modality influences
the d′ of the participant in the masking task.
Robinson et al. (2016) were interested in whether olfaction
interacts with vision. To investigate this topic, they used
object-substitution masking. In the study, participants were
presented with images such as pictures of a leaf of mint, an
orange, or a rose flower. These were called “target” images
because, in some conditions, they would be accompa-
nied by the odors associated with each plant. Participants
were also presented with “nontarget images,” which were
images, such as those of a hat or a telephone, for which cor-
responding odors would not be presented. The odor that
was presented just before each trial either corresponded to
a target (i.e., the smell of a rose before an image of a rose
was flashed) or did not correspond to the target (i.e., the
smell of mint was presented just before an image of a rose
was flashed). In some conditions, the stimulus was masked
with the object-substitution masking, whereas in other
conditions, the target and nontarget images were presented
ISLE 2.12
Masking Demonstration

45 Chapter 2: Research Methodology
without masks and could be more easily seen. The ques-
tion that Robinson et al. were interested in was whether or
not the presence of a corresponding odor would lower the
threshold (or raise the sensitivity) of detecting masked stim-
uli. See Figure 2.19 for an illustration of their methodology.
Here is a summary of Robinson and colleagues’ (2016)
results. First, the masking manipulation worked. For
the masked trials, sensitivity was higher than for the
unmasked trials. On catch trials (where no target was
present), participants mostly claimed they could not see
anything. Thus, these results suggest that the object-
substitution masking manipulation was successful at
reducing the visibility of masked trials. Having established
that the masking worked, we can now look at whether
the presence of odors influenced performance on the
visual task. Note that if odors either strengthen the effect
of the masking or make the participants more sensitive to
the target stimuli, we can say that there is an odor–vision
interaction. Only if the odors have no effect on masking
can we reject the idea of an odor–vision interaction, at
least in this type of study. The results here are mixed. For
men, there was no effect of odor on the masking variable.
Men showed the same sensitivity in conditions in which a
congruent odor (mint–mint), an incongruent odor (mint–
orange), or no odor was presented. However, there was a
very strong effect with women. Women showed a much
stronger sensitivity to targets when there was a congruent
odor, but also showed a weakened sensitivity to targets
when the odor was incongruent relative to the no-odor
condition. Thus, for women, smelling the same odor as
the masked stimulus made that stimulus easier to visually
detect. But also, for women, when they smelled an odor
that did not match the visual stimulus, detecting the stim-
ulus was harder.
These results are consistent with the general advantage
women have over men with respect to their sense of
smell (Zucco, Aiello, Turuani, & Koster, 2012). However,
beyond that, Robinson et al. (2016) show a greater inte-
gration between vision and olfaction in women than
in men. This greater integration helps women, relative
to men, when the odor and the image match, but hurts
women, relative to men, when there is a mismatch. If you
are interested, you can read Robinson’s paper to explore
neurological explanations for the effect. We present their
results here to show you that psychophysical methods
such as threshold detection, sensitivity, and the signal
detection approach are alive and well in the contempo-
rary field of sensation and perception.
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
MaleFemale
Congruent No Odor Incongruent
FIGURE 2.19 Sensitivity of Detection
The y-axis represents sensitivity. Lower values represent better
detection of the masked stimuli. Sensitivity in females was
helped by congruent odors, but sensitivity was also hurt by
incongruent odors. In men, however, odor had no effect at all,
either positive or negative.
APPLICATION: Psychophysics in
Assessment: Hearing Tests and Vision Tests
Our sensory capabilities are critical to living our lives in the
manner that we choose. One’s ability to make a livelihood,
attend school, and enjoy various recreational opportunities
is directly related to one’s ability to perceive the world accu-
rately. Take, for example, the everyday activity of driving a
car. Most of us have to drive to work or school. Driving
Masking: refers to the difficulty in seeing one stimulus when
it is quickly replaced by a second stimulus that occupies the
same or adjacent spatial locations

46 Sensation and Perception
is completely dependent on being visually competent. If
you become visually impaired, your ability to drive safely
decreases, and if you lose your eyesight, driving is no longer
an option. Once you get to school, you are also dependent
on your senses. You have to read a textbook and listen to
lectures. These require both vision and audition. Having
a conversation with anyone requires that your hearing
allow you to understand what others are saying. Even pain
allows us to recognize health issues that threaten us. Many
people will endorse the statement that eating delicious food
is one of life’s great pleasures.
But now, consider a student who is visually impaired or has
a hearing deficit. If you do not think you know any, you are
not paying attention. These impairments can range from
mild nearsightedness, requiring glasses or contact lenses,
to much more severe impairments. Being visually impaired
presents a number of obstacles that a sighted person cannot
even imagine. Even in a technological world that can offer
the visually impaired many ways of compensating for their
deficit and competing against normally sighted individuals,
being blind obviously creates obstacles. Blind students now
can use technology that automatically reads text aloud, but
text readers proceed more slowly than do sighted people
who typically can read while looking at text. Blind readers
who use Braille type can read silently, but this process is
also typically slower than reading by vision. So even with
the best technology, blind students work much harder and
longer to keep up with their sighted classmates. Auditory
impairment can be equally challenging. Those of us with
normal hearing take spoken language for granted. But deaf
individuals rely on sign language to communicate, which
renders reading and writing essentially the same as learning
a second language. Even less severe auditory impairment
can be challenging. Hearing aids may amplify sounds, but
it is often tricky to get them to amplify the sounds you want
to hear (e.g., voices) and not the sounds you do not want to
hear (e.g., traffic noise).
Assessing vision and hearing loss is an important task that
is fulfilled by professional optometrists and audiologists
with great competence. At the heart of each of these pro-
fessions is applying the ideas of psychophysics to diag-
nosing visual and auditory problems. In this Application
section, we take an in-depth look at this aspect of applied
psychophysics, that is, focusing on the psychophysics tests
optometrists and audiologists typically perform.
According to the National Institute on Deafness and Other
Communication Disorders, there are 35 million people in the
United States alone with some form of hearing impairment.
That’s roughly 10% of the nation’s population. In adults
65 years and older, the percentage exceeds 33% (National
Institutes of Health, n.d.). Hearing loss can be caused by
any number of factors, including genetic disposition, but
also age, disease, and exposure to noise. Hearing loss in
children and younger adults is usually due to genetic condi-
tions. But among older adults, hearing loss is just as likely
to be caused by environmental conditions. Hearing loss is
particularly common among those who work in professions
in which exposure to loud noise (or music) is chronic. This
includes airport employees, police officers, DJs and bartend-
ers at nightclubs, and, yes, rock musicians. Indeed, many
older rock musicians have serious hearing loss, including
Ozzy Osbourne, Phil Collins, and Pete Townshend.
Hearing loss can be divided into two broad categories,
sensorineural hearing loss and conductive hearing loss.
Sensorineural hearing loss refers to permanent hearing
loss caused by damage to the cochlea or auditory nerve.
This means that the hearing loss is due to problems in the
transduction of the physical sound waves into a neural
signal. Sensorineural hearing loss can result either from
genetic causes or from damage to hair cells or auditory
nerve fibers. Conductive hearing loss refers to the inabil-
ity of sound to be transmitted to the cochlea. This means
that the problem is that inadequate levels of sound reach
the cochlea to be transduced into a neural signal. Some
conductive hearing loss may be as simple as clogged path-
ways, but conductive hearing loss may also be permanent.
In many cases, medical treatment can restore hearing loss
caused by conductive causes, which may include earwax
buildup or a punctured eardrum (tympanic membrane).
Sensorineural hearing loss is seldom treatable, although in
extreme cases, cochlear implants can restore some hearing
lost through sensorineural damage.
The first step in assessing hearing loss is visiting a
professional audiologist. An audiologist is a trained
professional who specializes in diagnosing hearing
impairments. Audiologists usually train to the doctoral
level and obtain a degree known as a doctorate in audi-
ology (AuD). This qualifies them to diagnose and treat
hearing impairment (Figure 2.20). An audiologist can fit
a patient with the appropriate hearing aids or refer the
patient to a medical doctor if the audiologist thinks the
situation requires medical intervention.
Sensorineural hearing loss: permanent hearing loss caused
by damage to the cochlea or auditory nerve or the primary
auditory cortex
Conductive hearing loss: the inability of sound to be transmitted
to the cochlea
Audiologist: a trained professional who specializes in diagnosing
hearing impairments

47 Chapter 2: Research Methodology
The audiologist uses a device called an audiometer to assess
hearing loss. An audiometer can present tones of different
frequencies, from low in pitch to high in pitch, at different
volumes from soft to loud. A patient will listen on head-
phones in a soundproof room to a series of tones presented
by the audiometer. Tones are presented to one ear at a time, so
as to determine if there is hearing loss in each ear. Typically, an
audiologist will use the method of constant stimuli combined
with a touch of the forced-choice method and vary the loud-
ness of the tone among trials. The patient will typically indi-
cate if he or she heard the tone in the left or the right ear. If the
patient cannot hear the tone, this will be obvious from chance
performance for that frequency at that sound level. Usually,
the audiologist includes catch trials in which tones are not
presented. This allows the audiologist to determine the crite-
rion at which the patient is indicating hearing a sound.
After the test is complete, the audiologist will plot the
threshold for each frequency as a function of decibels.
That is, the audiologist makes a graph indicating how
loud each frequency needs to be in order for the patient
to hear it. The audiogram is the graph that illustrates the
threshold for the frequencies as measured by the audiom-
eter (Figure 2.21). The y-axis represents intensity, mea-
sured in decibels, and the x-axis represents frequency,
©
iStockphoto.com
/Fotosm
urf03
FIGURE 2.20 Audiologists
Audiologists are trained to treat hearing loss. They assess
hearing loss and help fit people with hearing aids. If medical
issues arise, they will forward patients to medical doctors who
specialize in auditory issues.
FIGURE 2.21 Audiology Report
This audiology report concerns a middle-aged woman with moderate sensorineural hearing loss. In the left graph, the y-axis indicates
how loud in decibel units a tone must be presented in order for the patient to hear it. The x-axis indicates frequency, from low
frequency (low notes) to high frequency (high notes). This patient shows maximum hearing loss for medium frequencies, unfortunately
in the range of human voices.

48 Sensation and Perception
measured in Hertz. The threshold is then compared with
a standardized curve, which represents normal or aver-
age hearing. This graph can then be used to determine
if there is hearing loss at any frequency. If there is, the
audiologist can help the patient with choosing hearing
aids or other remedies.
Particularly important for human hearing is the ability
to hear human speech sounds. Many hearing-impaired
people can detect sounds but have difficulty hearing
words, especially if there is background noise. A deficit
in hearing the voices of family members is often what
brings a patient to an audiologist in the first place.
Thus, audiologists will assess patients’ ability to hear
speech sounds against background noise. Specialized
tests have been developed to allow audiologists to check
for speech recognition in hearing-impaired individuals
in many different languages (Zokoll, Wagener, Brand,
Buschermöhle, & Kollmeier, 2012). In setting up hearing
aids for patients, audiologists can program the hearing
aids to filter out sounds that are not likely to be voices
and amplify sounds that are likely to be human voices.
When programmed correctly, hearing aids can selectively
filter out noise and increase the volume of human voices.
Turning now to vision, most of us may be more expe-
rienced with optometry, as minor visual problems are
more frequently corrected with eyeglasses than minor
auditory problems are corrected with hearing aids.
Optometrists have a similar job to audiologists but with
respect to vision. Optometrists provide most people’s
primary eye care. Optometrists have doctoral degrees in
optometry and are licensed to diagnose and give pre-
scriptions for many eye-related problems. But the most
common task they do is an eye examination, in which
they run a patient through a battery of tests, some to
look for diseases of the eye and others to look at basic
psychophysical properties of the vision of each patient.
Like audiologists, if an optometrist detects a medical
problem, that patient will be directed to an ophthalmol-
ogist, that is, a medical doctor who specializes in care
and diseases of the eye.
Included in the eye exam is a test of visual acuity. The test of
visual acuity measures a person’s ability to resolve an object
in focus at a particular distance. An individual person’s acu-
ity is then compared with a “standard” or normal reference.
The most common form of the test of visual acuity is the
ubiquitous Snellen chart (Figure 2.22). Using a Snellen chart,
an optometrist will ask a patient to read off the chart each
line until the letters become difficult to resolve. Typically,
this is done one eye at a time, as one eye may be normal
but the other may need to be corrected. Results from the
Snellen chart yield the typical measure of visual acuity that
most people are familiar with. A person of normal vision
is said to have 20/20 vision, as he or she can see what a
normal person can distinguish at 20 feet. If a person sees at
20 feet what a normal person sees at 60 feet, then that per-
son has 20/60 vision. If so, the optometrist would prescribe
corrective lenses, either glasses or contact lenses. In other
countries, these tests are done in meters, so 6/6 would be the
standard for 6 meters (6 meters = 19.7 feet).
Someone who has trouble seeing distant objects is
said to have myopia. Myopia is a condition in which
Audiometer: a device that can present tones of different
frequencies, from low in pitch to high in pitch, at different
volumes from soft to loud
Audiogram: a graph that illustrates the thresholds for the
frequencies as measured by the audiometer
Optometrist: a trained professional who specializes in
diagnosing visual impairments and diseases
©
iS
to
ck
ph
ot
o.
co
m
/a
lfa
br
av
oa
lfa
ro
m
eo
FIGURE 2.22 The Snellen Chart
A patient is asked to read the lowest line that he or she can.
This measures visual acuity and is the first test for myopia
(nearsightedness). Normal vision is considered 20/20. Anyone
who cannot see what a normal person can see at 20 feet may be
myopic.

49 Chapter 2: Research Methodology
incoming light does not focus directly on the ret-
ina but in front of it. An optometrist can easily treat
myopia by prescribing glasses or contact lenses. In
older adults, presbyopia becomes more frequent.
Presb yopia is a condition in which incoming light
focuses behind the retina, leading to difficulty focusing
on close-up objects. People with presbyopia need
glasses to read small print. Even adults who have
had 20/20 vision their entire lives are likely to develop
presbyopia as they age into their 40s and 50s. To see
what it would be like to have these conditions, go to
ISLE 2.13.
Optometrists serve as initial screeners and diagnos-
ticians of a number of diseases that affect the eye,
including glaucoma, macular degeneration, diabetic ret-
inopathy, and conjunctivitis. However, these diagnoses
do not require psychophysical testing, and therefore we
do not consider them
here. The American
Optometric Association
has valuable educa-
tional resources on its
website. If you are interested, go to http://www.aoa.org/
patients-and-public/eye-and-vision-problems?sso=y.
ISLE 2.13
Seeing With Myopia and Presbyopia
CHAPTER SUMMARY
2.1
Explain the nature of psychophysical scales and
how they measure the relations between real-
world stimuli and our perceptions of them.
In this chapter, we covered the basics of psychophysics.
The method of limits, the method of constant stimuli, and the
method of adjustment are used to determine psychophysical
properties of the observer, including the observer’s thresh-
olds, or ability to detect differences between stimuli along a
particular dimension. Absolute thresholds are the smallest
amount of a stimulus necessary for an observer to detect its
presence, whereas a difference threshold (also known as a
JND) is the smallest difference between two stimuli that can
be detected. Closely and inversely related to the concept of
threshold is the idea of sensitivity. Sensitivity is the ability to
perceive a stimulus. Magnitude estimation is a psychophysi-
cal method in which participants judge and assign numerical
estimates to the perceived strength of a stimulus. Magnitude
estimation fits Stevens’s power law, a mathematical formula
that describes the relation between the intensity of physical
stimuli and the magnitude of the perceptual response
2.2
Demonstrate an understanding of signal detec-
tion theory.
Signal detection theory is the theory that in every sensory
detection or discrimination, there is both sensory sensitivity
to the stimulus and a criterion used to make a cognitive deci-
sion. In signal detection theory, there is a trade-off between
successful detection of the target (e.g., a hit) and a detec-
tion response when there is no stimulus (e.g., a false alarm).
An observer wants to maximize his or her hits and correct
rejections (saying “no stimulus” when none is present) and
minimize his or her false alarms and misses (saying “no stim-
ulus” when a stimulus is actually present). Critical in signal
detection theory is the idea of sensitivity but also criterion,
an internal threshold, determined by the observer, above
which the observer makes one response and below which
the observer makes another response. Mathematically using
signal detection theory allows researchers to determine the
ROC curve, which is usually a plot of hits as a function of
false alarms.
2.3
Evaluate neuroscience methods and what they
tell us about sensation and perception.
Neuroimaging techniques are technologies that permit
visual examination of living human brains. These techniques
allow scientists to correlate perception with brain activity.
Two goals of neuroimaging are to reveal where perception
happens in the brain and how perception unfolds through
the brain over time. A number of different neuroimaging
techniques are reviewed in this chapter. EEG measures
electrical output of the brain. Although EEG provides good
temporal resolution of the brain, it lacks the spatial resolu-
tion of other methods. MEG measures the brain’s magnetic
fields. MEG provides good temporal resolution and better
spatial resolution than does EEG. In fMRI, magnetic fields
create a three-dimensional image that can capture both
the structure and the function of the brain. This technique
allows for good imaging of the brain in both time and space.
Finally, TMS is a procedure in which a magnetic coil is used
to stimulate electrically a specific region of the brain. This
allows one to probe the brain to see which areas produce
which kinds of perceptions.

Sensation and Perception50
REVIEW QUESTIONS
1. What does the Scoville scale measure? Why is it
considered to be a psychophysical scale?
2. What is the method of limits? How is it used to deter-
mine absolute thresholds?
3. What is the method of adjustment? How is it used to
determine the point of subjective equality?
4. What is the two-point touch threshold? How does it
illustrate the concept of a JND? How do two-point
touch thresholds differ across the human body?
5. What is the difference between response expansion
and response compression? How do both relate to
Stevens’s power law?
6. What is signal detection theory? How is it used to
predict performance on perception tests?
7. Define the terms criterion and sensitivity. How do
they interact in signal detection theory?
8. What are the neuroimaging techniques? How does
each one allow the brain to be examined in terms of
both space and time?
9. What is visual masking? How is it affected by consis-
tent and inconsistent odors?
10. How is an audiogram used to assess hearing loss?
How might an audiogram help an audiologist pro-
gram a hearing aid?
PONDER FURTHER
1. Why are the methods of psychophysics still relevant
and important when we have neuroimaging tech-
niques at our disposal? What information might we
still gather from psychophysics that we would not be
able to know by observing neuroimaging results?
2. Signal detection theory applies some basic mathe-
matics to understanding perceptual thresholds. But
we have also seen how it might apply in other situa-
tions, from military applications to deciding whether
your car needs to go to the shop. What other domains
of psychology or life in general might signal detection
analyses be good for? If you cannot think of any, try to
apply signal detection analysis to surfing!
KEY TERMS
Absolute threshold, 30
Ascending series, 31
Audiogram, 47
Audiologist, 46
Audiometer, 47
Capsaicin, 27
Catch trial, 36
Conductive hearing loss, 46
Correct rejection, 37
Criterion, 37
Crossover point, 31
d′ (d-prime), 41
Descending series, 31
Difference threshold (JND), 30
Electroencephalography (EEG), 43
False alarm, 37
Forced-choice method, 36
Hit, 37
Magnetoencephalography (MEG), 43
Magnitude estimation, 34
Masking, 44
Method of adjustment, 33
Method of constant stimuli, 32
Method of limits, 29
Miss, 37
Optometrist, 48
Point of subjective equality (PSE), 33
Psychophysical scale, 28
Receiver-operating
characteristic (ROC) curve, 41
Response compression, 34
Response expansion, 34
Scoville scale, 27
Sensitivity, 34
Sensitivity (signal
detection theory), 41
Sensorineural hearing loss, 46
Signal detection theory, 36
Stevens’s power law, 35
Transmagnetic
stimulation (TMS), 43
Two-point touch threshold, 31

Chapter 2: Research Methodology 51
Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e
SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review,
study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning.
Learning Objectives Digital Resources
2.1 Explain the nature of psychophysical scales and how they
measure the relations between real-world stimuli and our
perceptions of them.
Differences in Our Sensory Worlds: Invalid Comparisons
With Labeled Scales
The Two-Point Threshold: Not a Measure of Tactile Spatial
Resolution
2.2 Demonstrate an understanding of signal detection theory. “Utilizing” Signal Detection Theory
Tracking and Controlling Pain by Sight
What Is Signal Detection Theory?
2.3 Evaluate neuroscience methods and what they tell us about
sensation and perception.
Human Brain Activity Measured by fMRI
A Few Tips for Critically Evaluating fMRI Studies
Using EEG to Map Brain Dynamics

3Visual System: The Eye
LEARNING OBJECTIVES
3.1
Appraise the nature of light as electromagnetic energy to
explain the relation of wavelength to perception.
3.2 Sketch the anatomy of the eye to show how it makes vision possible.
3.3
Interpret how the retinae transduce light energy into
a neural signal that is sent to the brain.
3.4
Illustrate how the visual system adapts to the dark and then adapts
again to the light to help us adjust to changing light levels.
3.5
Diagram the receptive fields of retinal ganglion cells,
and show how they need contrast to operate.
3.6
Judge the different refractive errors and eye diseases
on how they affect the ability of the eye to see.
INTRODUCTION
We live in a visual world. As human beings, we typically first seek out information
around us with our amazing eyes and the visual system behind them. Indeed, our ability
to see affects so much of our lives. Before you read on, pause for a minute and think of
all the reasons for which we need to see. Write them down before reading on. We use
vision to read, to look at photographs on Instagram, and to ponder the stars at night by
observing them. We need to see to drive our cars, to cook our meals, to mow our lawns,
and to sew our clothes. Despite all the ways seeing is vital to us, most of us often take
seeing for granted. But all it takes is one scare with an eye injury, such as a detached
retina, or the onset of a disease, such as glaucoma, to make us realize how critical our
vision is to our everyday well-being. One of the authors (JHK) was playing pickup bas-
ketball. At his height, he is sometimes below the notice of defenders. In one game the
defender reached out and his fingernail scratched the cornea of the author’s eye. It was
a minor and temporary injury, thankfully, but for days the author was very sensitive to
light and had to wear sunglasses inside and often look down to avoid the lights. Driving
and many other visually demanding tasks were out of the question. Reading was nearly
impossible. It was a pretty tough way to begin the school year.
ISLE EXERCISES
3.1 The Basics of Waves
3.2 Accommodation
3.3 Presbyopia
3.4 Letters and the Fovea
3.5 Map Your Blind Spot
3.6 Photopic vs. Scotopic
Vision
3.7 Purkinje Shift
3.8 Convergence
3.9 Dark Adaptation
Function
3.10 Automatic Light
Adjustment in Cockpits
3.11 Center-Surround
Receptive Fields
3.12 Mach Bands
3.13 Correcting Myopia
and Hyperopia
3.14 Astigmatism
3.15 Living With Macular
Degeneration and
Retinitis Pigmentosa
3.16 Compound Eyes
3.17 Vision Prostheses
Mike Agliolo/Science Source

54 Sensation and Perception
Vision can also be a source of much joy to us (Figure 3.1). Imagine looking across
a field at a bright red rose in full bloom against the green foliage of its leaves. The
texture, color, and contrast all create an experience of beauty. Consider further being
at the Louvre museum in Paris, patiently waiting for your chance to see the Mona
Lisa. You’ve paid thousands of dollars and traveled thousands of miles to wait
uncomfortably in a packed crowd, just to get a glimpse of this famous Leonardo
da Vinci painting, a distinctly visual experience. Think about watching a movie and
how much excitement and relaxation you may get from watching, that is, seeing,
the adventure unfold before you. And perhaps imagine looking at the face of your
significant other; the sight of this person’s face should also bring you joy.
To summarize, our visual system provides us with necessary information all the
time, so we can find our way, not bump into walls, and obtain information from the
landscape. It also transcends these practical aspects by providing us with a sense of
beauty and wonder. Whether mundane or majestic, most of all visual experiences
FIGURE 3.1 Visual Images Can Be Beautiful
We often find great beauty and emotional comfort in merely looking at visual images. Do these
images make you smile?
©
S
hu
tt
er
st
oc
k/
Sc
an
ra
il1
©
S
hu
tt
er
st
oc
k/
ca
na
da
st
oc
k
©
S
hu
tt
er
st
oc
k.
co
m
/S
eb
as
tia
n
D
ud
a
(a) (b)
(c) (d)
©
S
hu
tt
er
st
oc
k.
co
m
/M
en
no
S
ch
ae
fe
r
(f)
©
iS
to
ck
ph
ot
o.
co
m
/L
eo
Pa
tr
iz
i
(e)
Sh
ut
te
rs
to
ck
.c
om
/O
lg
a
Sa
vi
na

55 Chapter 3: Visual System: The Eye
start with a light source, which is where this chapter starts. Before we can talk about
any part of the visual system, we need to introduce light as a scientific concept.
Review the sequence of steps of sensory systems in ISLE 1.1. We are starting at the
beginning of this sequence when we talk about light.
LIGHT
3.1
Appraise the nature of light as electromagnetic energy to
explain the relation of wavelength to perception.
Vision is the sensory system that allows us to perceive light. Thus, before we begin
describing the visual system, it is important to know just a little bit about the nature
of light. We live in a world bathed in light, sunshine by day, artificial light by night,
and the constant omnipresence of glowing screens, big and small, but what exactly
is light? Visible light is an example of electromagnetic energy (Feynman, 1985,
1995). As most of us know, light, like other forms of electromagnetic radiation,
moves very fast—approximately 186,000 miles per second (300 million meters per
second) in a vacuum. Visible light is one part of the electromagnetic spectrum, which
also includes radio waves, infrared radiation, ultraviolet radiation, and gamma rays.
These varied types of electromagnetic radiation differ only in their wavelengths,
which we discuss shortly. According to modern physics, light, curiously, is made up
of particles called photons that also behave in a wavelike manner (see Feynman,
1985, for a more thorough explanation of our current understanding of light in
lay terms). The wavelike behavior of light is important to vision, as we see differ-
ent wavelengths (or frequencies, wavelengths’ inverse) as distinct colors. So, in this
chapter, we are concerned with light behaving more as a wave phenomenon than as
discrete photons.
When thinking of light as a wavelike phenomenon, we can look at other waves
to help us understand light. Waves are waves. So, consider waves at the beach
before they break on the shore. The waves are up-and-down undulations of the
water’s surface. As a result, waves are usually drawn as going up and down (ISLE
3.1). If you watch the waves coming in, you can tell that not all waves are the
same. Some are taller than others, and some peaks are closer together, and other
peaks are farther apart. Surfers look for “sets” of waves that come in with higher
peaks, though these higher peaked waves also often come close together, so that a
surfer may be able to ride only one per set. These observations allow us a way to
describe what makes waves different from each other. We can quantify these dif-
ferences by both how high they are (amplitude) and how close they are to adjacent
waves (wavelength). Thus, the distance between the peaks of a wave is called the
wavelength, and the height of the wave is called its intensity. We call the amplitude
of a wave intensity because the higher the wave is, the stronger it is. Think of the
difference between ripples (low amplitude), the waves you might bodysurf on a
day on the beach (medium amplitude), and a tsunami, the ultimate high-amplitude
wave. The number of waves per unit of time is called the frequency, which is equal
to 1 divided by the wavelength. In perceptual terms, wavelength roughly correlates
with color, and intensity correlates with brightness. To make matters confusing for
students everywhere, the tradition is to refer to light waves by their wavelengths
but sound waves by their frequencies. Because this is the tradition throughout the
field, we respect it here.
Electromagnetic energy:
a form of energy that includes
light that is simultaneously both
a wave and a particle
Wavelength: the distance
between two adjacent peaks
in a repeating wave; different
forms of electromagnetic
energy are classified by their
wavelengths
Intensity: when referring to
waves, the height of a wave
Frequency: the number
of waves per unit of time;
frequency is the inverse of
wavelength
ISLE 3.1
The Basics of Waves

56 Sensation and Perception
Visible light represents a tiny slice
of the electromagnetic spectrum. The
electromagnetic spectrum is the com-
plete range of wavelengths of light
and other electromagnetic energy. We
measure the wavelength of light in
nanometers (nm). A nanometer is one
billionth of a meter. Visible light rep-
resents wavelengths between 400 and
700 nm, but electromagnetic radiation
extends in both directions beyond those
boundaries, from wavelengths being far
shorter to wavelengths of several miles
(Figure 3.2). Wavelength is also related
to energy. The shorter the wavelength
is, the higher the energy is. The longer
the wavelength is, the lower the energy
is. Thus, wavelengths shorter than visi-
ble light come with higher energy, such
as gamma rays, X-rays, and ultravio-
let rays. With longer wavelengths than
those of visible light comes lower energy
radiation, such as infrared radiation, microwaves, and radio waves.
Consider an old-fashioned candle (Figure 3.3). The burning wick creates electro-
magnetic radiation at many wavelengths. Some of the electromagnetic radiation is in
the visible wavelength range. We see this radiation as the pretty yellow glow of the
flame, with the (higher energy) blue by the wick itself and the less energetic yellow light
toward the top. We also feel some of the energy in the form of heat, which is electro-
magnetic energy being emitted in the infrared range.
The shortest wavelengths we see are about 400 nm, and we perceive these as
violet. Blue has a wavelength of about 470 nm, and green has a wavelength of about
500 nm. Red has a wavelength approaching 700 nm, the longest wavelengths that we
can perceive under normal circumstances. Other animals have different sensitivities
to wavelength of light. Bees, for example, can see shorter wavelengths than we can
(ultraviolet) (Horridge, 2012). Flowers are colorful to us, but bees and hummingbirds
may use ultraviolet coloring of flowers to determine when they can obtain nectar
from those flowers. Some birds use ultraviolet coloration to attract mates. For exam-
ple, male blue tits have ultraviolet crowns on their heads, which they display during
mating courtships. A human observing the birds would not see any differences, but
female blue tits prefer shinier ultraviolet colorings (Figure 3.4). Interestingly, these
ultraviolet colors are not visible to predators who tend to see better in longer wave-
lengths, allowing the male to show off without drawing too much attention from
predators (Håstad, Victorsson, & Ödeen, 2005). Even some mammals such as bats
and rodents see shorter wavelengths than we do (Beran, 2012).
Recall that in physics terms, light can be thought of as either a wave phenomenon
or a particle phenomenon. When physicists are thinking of light as a wave, they refer to
wavelength, but when they are thinking of light as a particle phenomenon, they refer to
photons. Photons are single particles of light. Photons are useful for thinking about the
amount of light. A very bright light has many more photons than a dim light, regard-
less of its wavelength. This means that bright sunshine sends many more photons your
way than does a single candle burning at night. As psychophysicists, we use the ter-
minology of wavelength when discussing color but the terminology of photons when
400 nm
400 500
Visible
light
Gamma rays
10–11 10–9 10–7 10–5 10–3 10–1 101 103 105 107 109 1011 1013 1015 1017
Radio waves
x
rays
UV
rays
Infrared
radiation
Wavelength (nm)
Micro-
waves
600 700
500 nm 600 nm 700 nm
FIGURE 3.2 The Electromagnetic Spectrum
Electromagnetic radiation varies in wavelength from very low-energy,
long-wavelength waves (shown to the right of the visible spectrum) to very high-
energy, short-wavelength waves (shown to the left of the visible spectrum). Visible
light is just a tiny segment of this continuum. Within the visible spectrum, we see
shorter wavelengths as bluish and longer wavelengths as reddish.
Electromagnetic spectrum:
the complete range of
wavelengths of light and other
electromagnetic energy
Photon: a single particle
of light

57 Chapter 3: Visual System: The Eye
discussing intensity and brightness. We also use the concept of light as particles when
talking about the process of transduction, the process whereby the eye captures light.
Most light reaches our eyes indirectly, after bouncing off objects in our world.
We see the rose in the garden because its surface reflects the light of the sun (or your
flashlight, if you are prowling around the garden after dark). Surfaces, such as the
rose petals, tend to absorb some wavelengths and reflect others. We see the rose as
red because it absorbs most wavelengths but reflects the long wavelengths that we
perceive as red. Similarly, the grass is green because the plant is absorbing most
wavelengths, but reflecting green light back into the world.
TEST YOUR KNOWLEDGE
1. Evaluate the nature of electromagnetic radiation, focusing on the range that is considered
light.
2. Create a diagram to show how the selective absorption and reflection of different
wavelengths allows us to see different colors.
THE EYE AND ITS ROLE
IN THE VISUAL SYSTEM
3.2 Sketch the anatomy of the eye to show how it makes vision possible.
It will not come as a surprise to most readers that the eye is the primary organ of visual
perception, what was called the accessory structure in ISLE 1.1. However, how the eye
converts light into signals that allow us to see colors, objects, motion, and a third dimen-
sion is a fascinating topic, which this and the next several chapters cover. The eye is a
finely tuned instrument, with many integrated parts working together to capture light
in a way to allow our visual perception. To understand visual perception, it is necessary
to understand the anatomy and the physiology of the eye. Thus, our first task will be
to consider the structure of the human eye, its physiology, and how it relates to seeing.
The human eye is a fairly typical mammalian eye. See the Exploration section for
a discussion of a few other types of animal eyes. Like most primate eyes, our eyes are
located at the front of our heads and are slightly spaced. This gives us good depth per-
ception. Relative to other mammals, human beings have excellent visual acuity (i.e.,
the ability to see fine details) but are relatively poor at seeing in dim light compared
with familiar animals such as cats and dogs. Cats are especially good at seeing in dim
light. Because of the structure of the receptors in the retina, humans are good at color
vision. Dogs can see in color but are similar to humans with red–green color defi-
ciency. Other animals seem to have better color vision than us, though. Indeed, gold-
fish see many more colors than we do (Neumeyer, 1985). It is interesting to think that
compared with your pet goldfish, swimming in its bowl, you are seemingly colorblind.
The basic story of how humans see is as follows. Light emanates from a light
source, such as the sun or an overhead light in an indoor location. Light falls on
objects in the environment, and, in turn, some wavelengths are reflected by those
objects. It is this reflected light from objects that we actually see. The reflected light
enters the eye through the pupil and is focused on the retina by the cornea and lens.
The retina contains specialized cells called rods and cones, which transduce, or con-
vert, the light energy into an electrochemical signal, which is then sent to the brain for
processing through the optic nerve. Of course, humans, like most animals, have two
FIGURE 3.3
Light in the Visible Spectrum
A burning candle emits light in
the visible spectrum. We can
use this light to illuminate a room
when the electricity is off and we
cannot use our electric lights.
©
iStockphoto.com
/aga7ta
FIGURE 3.4 Blue Tits
Blue tits are pretty birds, with
beautiful colorations. However,
there are colors present in the
ultraviolet range that we cannot
see, but other blue tits can. Blue
tit females prefer males with
shiny ultraviolet crowns.
©
iStockphoto.com
/G
lobalP

58 Sensation and Perception
Visible
60°
80°
Invisible
Invisible
Invisible
Visible
to both eyes
(stereoscopic
depth
perception
possible)
Visible to
right
eye only
Visible to
left
eye only
110° 190°
Invisible
Invisible
Invisible
FIGURE 3.5 Field of View
Our eyes give us a large zone of binocular vision, that is, an area we can see with both eyes. Beyond that is a region in which
we see with only one of our eyes. However, we cannot see the world behind us. If you think of the world as a circle around
you, you can see approximately 190 degrees in front of you, but you cannot see the 170 degrees behind you.
Field of view: the part of the
world you can see without eye
movements
eyes, so coordinating the slightly different view of each eye is critical to vision, espe-
cially depth perception. So, let’s look at each of the events in greater detail.
Field of View
Human eyes are located approximately 6 cm (2.4 inches) apart from each other. Of course,
there is variation depending on genetics and how big one’s head is. In humans, as in other
primates, the eyes have evolved to be adjacent to each other to allow for accurate depth
perception. Frontally placed eyes allow us to see in stereo, as each eye has a slightly differ-
ent view of the world. By computationally comparing the two images, our visual systems
can estimate distance from the offset of the two images. Such depth perception is useful
for animals that must hunt for food or that live in trees and must judge distances from
branch to branch. The drawback of this arrangement is that such animals can see only in
front of themselves (Figure 3.5). Large herbivores, such as deer and antelope, are often
subject to predation. For these animals, detecting incoming predators is paramount. For
this reason, these species have evolved eyes placed on the sides of the head rather than
in front. Animals with eyes positioned on the sides of their head such as antelope have a
greater field of view but weaker depth perception (Figure 3.6). This allows them to detect
predators coming from almost any position relative to them, except usually a small area
directly behind their heads. Your field of view is the part of the world you can see without
eye movements. For humans, that is approximately 190 degrees horizontally (side to side)
and 140 degrees vertically (up and down).
Anatomy of the Eye
The complexity of the human eye has often led to debates as to how it got that way. Before
Darwin, the eye was often held up as a marvel of intelligent design. Indeed, the naturalist
William Paley labeled the design of the eye a “miracle.” For Paley, he could not see how
such a complex piece of biological machinery could have just happened. He reasoned that
because it was so complex, it must have been designed. Perhaps because of this challenge,
many evolutionary biologists have studied the evolution of complex eyes in great detail.
Biologists have carefully mapped out how nature has selected for complex eyes and have
FIGURE 3.6
Pronghorn Antelope
The pronghorn antelope can be
found in the American West,
such as this one in Montana. The
pronghorn antelope has eyes on
the sides of its head. This limits its
depth perception but allows it to
see behind its head as well as in
front. This allows the pronghorn
antelope to detect incoming
predators, such as mountain lions.
©
iS
to
ck
ph
ot
o.
co
m
/p
ho
to
gr
ap
hy
by
JH
W
ill
ia
m
s

59 Chapter 3: Visual System: The Eye
investigated various intermediaries. Indeed, complex eyes have
evolved independently. For example, the octopus is a mollusk
more related to clams and barnacles than it is to us. However,
the octopus has a complex eye that evolved along a very differ-
ent pathway than our human eye. Our concern here is not how
eyes evolved in the distant past but how humans use their eyes
now. The result of human evolution is a remarkably complex
apparatus designed for transforming light energy into percep-
tual information.
In this section, we describe the anatomy and physiology
of the parts of the eye, with an emphasis on their func-
tion in vision (review Figure 3.8 first). Make sure you learn
both the vocabulary and the concepts. That is, memorize
the technical terms, but also make sure you understand
the function of each part of the eye and how information
moves from the eye to the brain. We start now with the
cornea and iris, the outermost parts of the eye.
The Cornea
The cornea is the clear front surface of the eye that allows
light in. It is also a major focusing element of the eye. The
cornea is transparent and is the outermost surface of the eye
(see Figure 3.8). It begins the process of refracting or bending
the light to come into focus on the retina of the eye, which
Cornea: the clear front surface
of the eye that allows light in;
also a major focusing element
of the eye
Optic
axis
Pupil Cornea
Iris
Lens
Sclera
Choroid
Retina
Fovea
Sheath
Optic nerve
Optic disc
(blind spot)
Vitreous
chamber
Zonule
fibers
Ciliary muscle
Posterior chamber
Anterior chamber
FIGURE 3.8 Anatomy of the Human Eye
Light enters the eye through the pupil, an opening in the outer layer of the eye, surrounded by the iris,
the colored area that gives eyes their characteristic color (brown, blue, green, etc.). Ciliary muscles
control the zonule fibers that pull on the lens, allowing light to focus on the retina at the back of the eye.
The fovea is a little pit at the center of the retina that picks up light from the point in space the eye is
looking at. Information leaves the eye via the optic nerve through a part of the eye called the optic disc.
FIGURE 3.7 Retina Layers, Colored Scanning
Electron Micrograph (SEM)
The retina of the human eye is found on the inside of the eyeball
and is formed from a number of layers. The light-detecting part
is formed of rod and cone photoreceptor cells. The lower third
of this image contains the blood vessels and nerves that form
the upper layer. The light entering our eyes has to penetrate
this layer to reach the tightly packed photosensitive “outer
segments” of the photoreceptors (top third of image) that lie
just below a layer of retinal pigment epithelium cells (top).
Magnification: ×480 when printed at 10 centimeters across.
Louise H
ughes/Science Source
ISLE 3.2
Accommodation
ISLE 3.3
Presbyopia

60 Sensation and Perception
is the back surface inside the eye. Indeed, the cornea does
most of the focusing of images toward the retina. However,
the cornea is rigid and can only focus clearly at one dis-
tance. Therefore, changes in refraction to allow us to see
clearly at different distances must come from the adjust-
able lens. The cornea is the transparent part of the sclera, a
tough membrane, which provides a protective covering for
the eye. The sclera is the “white” of the eye as well. Between
the cornea and the iris is the fluid-filled anterior chamber.
Just behind the cornea is a part of the eye called the iris.
The iris is really two muscles with an opening in the mid-
dle. Light enters through this opening. The opening itself
is called the pupil. In dim light, the iris opens to allow the
pupil to expand (or dilate) and it allows more light into
the eye. In bright light, the iris contracts and the pupil
narrows, allowing less light into the eye. This is a process
that is not directly under conscious control and is known
as the pupillary reflex. Shine a bright light into someone’s
eyes, and his or her pupil will immediately contract. When
you enter a dim room after being outside in the sunshine,
your pupils will quickly dilate. The two eyes coordinate
their pupillary reflexes, so that each dilates and contracts
simultaneously. In this way, the pupil can range from a
diameter of about 2 mm to about 8 mm, depending on
lighting conditions (Figure 3.9). Interestingly, it is the iris that gives individual eyes (or
pairs of eyes) their characteristic colors, whereas the pupil appears dark or black (Figure
3.10). The amount of melanin in the iris is a major factor in whether one has blue eyes
or brown eyes, though other factors contribute as well. Heterochromia is a condition
in which a person has two irises of different colors. It is rare in people (but not in
Siberian huskies) and is often a sign of ocular disease. The posterior chamber is the
space between the iris and the lens. It is filled with fluid, known as aqueous humor.
The Lens
The lens (also called the crystalline lens) is the adjustable focusing element of the eye. It
is located just behind the iris. This process of adjusting the focus for different distances by
changing the shape of the lens is called accommodation. Accommodation is the process of
FIGURE 3.9 The Pupillary Reflex
In dim light, the iris relaxes control, and the pupil expands (or dilates) to allow more light into
the eye (a). However, when a bright light is shined into the eye, the iris contracts and the pupil
narrows, allowing less light into the eye (b).
(a) (b)
FIGURE 3.10 Irises
The iris is the part of the eye that gives the eye its characteristic
color. Those eyes containing the most melanin will appear brown.
Those with less melanin may appear blue or sometimes green.
©
iS
to
ck
ph
ot
o.
co
m
/P
eo
pl
eI
m
ag
es
Sclera: the outside surface of
the eye; a protective membrane
covering the eye that gives the
eye its characteristic white
appearance
Anterior chamber: the fluid-
filled space between the cornea
and the iris
Iris: the colored part of the
eye; a muscle that controls the
amount of light entering through
the pupil
Pupil: an opening in the middle
of the iris
Pupillary reflex: an automatic
process by which the iris
contracts or relaxes in response
to the amount of light entering
the eye; the reflex controls the
size of the pupil
Heterochromia: a condition in
which a person has irises of two
different colors
Posterior chamber: the space
between the iris and the lens;
it is filled with fluid known as
aqueous humor

61 Chapter 3: Visual System: The Eye
adjusting the lens of the eye so that you can see both near and far objects
clearly. This process is very rapid, although changing accommodation
from a near object to a far object is faster than from a far object to
a near object (Kirchhof, 1950). Accommodation is controlled by mus-
cles connected to the lens, called ciliary muscles. The ciliary muscles
work in concert with the zonule fibers, which connect the lens to the
choroid membrane. As with the control of the iris, the ciliary muscles
work automatically, without conscious control. Indeed, we seldom need
to think about whether what we are looking at is in focus. The ciliary
muscles can contract and increase the curvature of the lens so that the
lens thickens. The increased curvature of the lens allows the eye to focus
on a close object. When the person then looks at a faraway object, the
muscles relax, and the focus of the lens changes to an object farther away
(Figure 3.11). For a demonstration of how this works, go to ISLE 3.2.
The process of accommodation allows us to focus on near objects
and far objects and switch between them effortlessly. Look up from
your textbook now and look out the window. As you do, your lens
relaxes and allows you to focus on a distant object. Look back at the
text and your lens contracts, allowing you to focus on the words on
the page or screen. But accommodation has limits. Close one eye and
with the other look at your index finger at arm’s length and then slowly
bring it toward your eye. At some distance, probably close to your
nose, you will no longer be able to focus on your finger, and it will start
to look a bit blurry. If you can bring your finger right up to your nose
and still focus on it, switch to a smaller object. When you can no longer
focus on the object, you have reached your near point. The near point
is the closest distance at which an eye can focus.
As people age, the near point shifts farther away from the eyes.
For younger adults, say, around 20 years of age, the near point is
usually about 10 cm from the eyes. However, it averages about 100
cm for people around 60 (Rutstein & Daum, 1998). This condi-
tion in older adults is called presbyopia. Because of the decreased
elasticity of the lens, accommodation does not occur to the extent
it does in younger eyes. This relative failure of accommodation
moves the near point farther from the eye. Presbyopia usually
starts to become noticeable in adults in their 40s and accelerates
after that (Figure 3.12). When presbyopia first becomes noticeable,
people may simply hold their reading material farther from their
eyes. But as it progresses, reading glasses or bifocals, which help
bring near images into focus, become necessary. For a demonstra-
tion of how this works, go to ISLE 3.3.
Development: The
Emerging and Aging Eye
The eye is an incredibly complicated structure. A natural question is
about how it develops prior to birth. This is not a text on embryol-
ogy, but we can give a brief overview of the incredible speed of the
development of the major structures of the eye. The eye begins to
develop very early after fertilization—by the 17th day in fact. That
is less than 3 weeks after conception. Shortly thereafter, the retina
and lens begin to develop. By the 32nd day you can identify the lens,
Retina
A
A
B
Lens
Cornea
Object far – eye relaxed
Focus on retina
Focus on retina
Focus behind retina
Object near – eye relaxed
(a)
(b)
Object near – accommodation(c)
Moving object
closer pushes
focus point back
Accommodation
brings focus
point forward
FIGURE 3.11 Curvature of the Lens
(a) The increased curvature of the lens allows the
eye to focus on a close object. (b) When the person
then looks at a faraway object, the muscles relax
and the focus of the lens changes to an object
farther away.
60
50
40
30
Near point of newborn
about 5 cm
Near point about 400
cm (4 m) by age 70
Near point about
10 cm by age 20
N
e
a
r
p
o
in
t
(c
m
)
Age (years)
20
10
0
0 10 20 30 40 50 60 70
Near point
about arm’s
length (40 cm)
by mid-50s
Near point about
reading distance
(20 cm) by mid-40s
FIGURE 3.12 Presbyopia and the Near Point
In this graph, the x-axis represents age, and the y-axis
represents the near point in centimeters from the eye. As
you can see in the graph, the near point is closest when
we are young and then accelerates after we reach the
age of approximately 50. There are very few among us
who do not need reading glasses by the time they are in
their 50s. This presbyopia (“old vision”) is the result of the
fact that the lens becomes less flexible as we age.

62 Sensation and Perception
and by day 35 you can start to see the iris. Although the retina, the most complex structure
of the eye, begins to develop early in utero, it continues to develop for months after birth
(Gilbert, 2000).
Turning now to aging, in the discussion of presbyopia, we described one way in
which the eye changes as we age. This is one of many changes, but let us examine one
other way that the lens changes as we age. Our lens loses its transparency to some degree
as we age, and not as much light gets through. However, it does not lose the ability to
transmit all wavelengths equally; rather, the lens loses the ability to transmit blue wave-
lengths the most (Salvi, Akhtar, & Currie, 2006). So, as we age, we are less able to see
blue colors. Figure 3.13 shows a simulation of how a scene may look to a person with
an older lens that extremely limits blue wavelengths. Figure 3.13a shows the scene as
normal, and Figure 3.13b shows the scene through the older lens. Figure
3.13b looks like a yellow filter has been placed over the scene.
TEST YOUR KNOWLEDGE
1. Sketch the eye and indicate how the cornea and lens work together to
produce a focused image on the retina. Then indicate how the lens changes
to allow us to see near and far objects clearly.
THE RETINA
3.3
Interpret how the retinae transduce light energy
into a neural signal that is sent to the brain.
In English, the word retina is singular, referring to one eye. The words reti­
nae and retinas are both acceptable as the plural form of the word retina. To
be more formal, we will use retinae when talking about both retinae. Before
we discuss the retina, perhaps the most complex part of a complex eye, let
us review what we have covered so far. The iris and the lens direct light and
focus that light. The pupil allows more or less light in, depending on ambient
conditions. The goal of these processes is to focus an image on the retina,
which is the eye’s photosensitive, that is, light-capturing, surface. This image
is known as the retinal image. Receptor cells on the retina transform the
light image into a neural signal to be sent to the brain. The intricacy and
precision of the arrangement of the retina is very complex but at the same
time fascinating (Figure 3.14). Although very thin, the retina is composed of
several interconnected layers. Let’s see how this process takes place.
FIGURE 3.14 An Optometrist’s View
of the Retina of One of the Authors
This is an image of the fundus, or the back
surface of the right retina. The bright yellow
spot is the optic disc, where the optic nerve
leaves the eye to head into the brain. This
also corresponds to the blind spot of the eye.
The dark patch just to the right of the yellow
disc is the macula or fovea, the part of the eye
designed for visual acuity. The eye depicted
belongs to Dr. Bennett Schwartz.
Lens: the adjustable focusing
element of the eye, located right
behind the iris; also called the
crystalline lens
Accommodation: the process
of adjusting the lens of the eye
so that both near and far objects
can be seen clearly
Ciliary muscles: the small
muscles that change the
curvature of the lens, allowing
accommodation
Zonule fibers: fibers that
connect the lens to the choroid
membrane
Near point: the closest
distance at which an eye can
focus
Presbyopia: a condition in
which incoming light focuses
behind the retina, leading to
difficulty focusing on close-up
objects; common in older
adults, in whom the lens
becomes less elastic
FIGURE 3.13 The Aging Lens and Loss of Seeing Blue
(a) You see a scene as normal, in effect, through a young lens. (b) You see the scene, somewhat
exaggerated, as through an older lens that does not let blue wavelengths through as well as other
wavelengths. The scene looks like a yellow filter has been place over the image.
(a) (b)

63 Chapter 3: Visual System: The Eye
Anatomy of the Retina
The retina is the third and innermost of the three membranes in the eye. Retinal anatomy is
quite complicated, but it can be understood when one thinks of the functions of the retina.
First, the retina is the location in the eye where transduction takes place. Photoreceptors
convert light into a neural signal. Thus, part of retinal anatomy is designed to help the
photoreceptors capture light. Second, the retina starts the process of transmitting visual
information to the brain. Thus, part of the retinal anatomy is designed to allow the retina
to send accurate signals to the brain. Figure 3.15 illustrates this complex anatomy, and
Table 3.1 lists the functions of different neural cells in the retina. Looking at the retina,
you can see that it is organized into layers. Let us look at the cells in these layers and see
how they connect to each other. Interestingly, some cells help carry the signals toward the
brain, whereas others seem to connect within their layers. The cells at the back of the eye
are the photoreceptors. These cells then synapse (i.e., connect) with two types of cells. The
horizontal cells connect from one photoreceptor to another while the bipolar cells convey
the signal from the photoreceptors to the next inner layer. The bipolar cells synapse with
amacrine cells and ganglion cells. Amacrine cells connect across different bipolar cells and
ganglion cells. The ganglion cells send their axons to the optic disc where they exit the
eye. These axons then become the optic nerve carrying the signals from the retina to the
brain. These bipolar cells and amacrine cells are interesting in that they connect within
their layers. Later on, we will discuss the important role these cells play in vision.
Ganglion cell layer
Inner synaptic layer
Inner nuclear layer
Outer synaptic layer
Outer nuclear layer
Inner and outer segments
of photoreceptors
Pigment epithelium
V
itr
eo
us
c
ha
m
be
r
Choroid
Pigment
epithelium
Axons of
optic nerve
Retinal
ganglion cell
Amacrine
cell
Bipolar
cell
Horizontal
cell
Cone Rod
C
ho
ro
id
Fovea
Fovea Retina
Retina
Ganglion cell layer
Retinal ganglion cells
Inner nuclear layer
Bipolar, horizontal, and
amacrine cells
Outer nuclear
layer
Photoreceptors
(rods and cones)
Inner and outer
segments of
photoreceptors
Inner synaptic layer Outer synaptic layer
(a)
(b)
FIGURE 3.15
Anatomy of the Human
Retina
In Figure 3.15a, we see the
location of the retina relative
to other anatomical structures
in the eye. As you can see, the
foveal region looks like a pit,
as the top layers of the retina
are swept aside so that there
is no scatter of light reaching
the receptor cells (i.e., the
cones). In Figure 3.15b, we
see a schematic of what the
retina would look like if it were
stretched out to reveal the
layers. Oddly, the rods and
cones form the back layer of
the retina, even though they
are the light-sensitive portion
of the retina. The horizontal,
amacrine, and bipolar cells
serve as a middle layer and
connect the photoreceptors
to the retinal ganglion cells,
which exit the retina and bring
visual information to the brain.
Retina: the paper-thin layer
of cells at the back of the eye
where transduction takes place
Retinal image: the light
projected onto the retina

64 Sensation and Perception
The Receptors: Rods and Cones
There are two types of receptors in vision (or photoreceptors), the rods and the cones.
These are the cells that transform light energy into an electrochemical signal. Rods predom-
inate at the periphery (away from the center) of the retina, whereas cones cluster in and near
the fovea of the retina. The fovea is a small pit at the center of the retina (indeed, the word
fovea means “small pit” in Latin). The fovea and the area around it are called the macula.
Beyond that is the periphery of the retina. There are about 120 million rods and only about
7 million cones in each eye. There is only one type of rod, whereas there are three classes of
cones (Nathans, Piantanida, Eddy, Shows, & Hogness, 1986; Neitz & Jacobs, 1986, 1990).
We return to these cone types a bit later in the chapter, as each is important in detecting
different frequencies of light and perceiving color. Because there is only one type of rod,
these receptors see only in shades of gray (from black to white). The rods and cones are not
spread evenly across the eye. If you examine Figure 3.16, you will see that the cones are
most centrally located in the fovea. There are no rods in the fovea. The rods are mostly at
the periphery, starting about 20 degrees off the fovea away from your nose and toward your
temples. In one region, however, the two receptor types mix. This is in the outer areas of the
fovea, or where the fovea blends into the periphery, sometimes called the parafoveal region.
Here, both kinds of cells can be found (Curcio, Sloan, Kalina, & Hendrickson, 1990).
The human fovea is densely packed with cones. It looks like a little pit on the ret-
ina because the cells that are above the retinal surface, such as retinal ganglion cells,
horizontal cells, and amacrine cells, are pushed to the side so that the cones are at the
surface making the retina even thinner. Because of the layers that are swept away, there
is less scattering of light in the fovea, allowing visual acuity to be higher in the fovea. It
is the foveae of the retinae that give humans our excellent color vision and our excel-
lent visual acuity. By visual acuity, we mean clarity of vision. Thus, the visual image
of the words you are reading right now is falling on the foveae of your retinae. Words
you have just read or have not yet read are tailing off into the periphery of the retinae.
When you move your eyes to look at something new, say, a family member walking
into the room, it is your foveae that are directed toward the new stimulus. Therefore, it
is said that the foveae represent the direction of your gaze. For a demonstration of how
much space on the retina is occupied by the high-acuity fovea, go to ISLE 3.4.
The fovea is unique anatomically. First, it is the location on the retina with the
highest density of cones. Second, other retinal cells are brushed away from the sur-
face of the retina. This allows light to reach the surface of the fovea with minimal
scatter of that light. Both of these factors evolved to enhance our visual acuity.
ISLE 3.5
Map Your Blind Spot
ISLE 3.4
Letters and the Fovea
Neuron Type Function
Receptors Transduce light into a neural signal
Rods Night vision, light detection, grayscale vision
Cones High visual acuity, color vision, daytime vision
Horizontal cells Receive information from photoreceptors and other horizontal cells;
cross talk across photoreceptors
Bipolar cells Receive information from photoreceptors; send signals to retinal
ganglion cells
Amacrine cells Receive information from bipolar cells and other amacrine cells; cross
talk function
Retinal ganglia Receive information from bipolar cells; send signal to brain via the
optic nerve
Rods: photoreceptors at the
periphery of the retina; they
are very light sensitive and
specialized for night vision
Cones: photoreceptors in the
fovea of the retina; they are
responsible for color vision and
our high visual acuity
Fovea: an area on the retina
that is dense in cones but lacks
rods; when we look directly at
an object, its image falls on the
fovea
Macula: the center of the
retina; the macula includes the
fovea but is larger than it
TABLE 3.1 Neurons in the Retina

65 Chapter 3: Visual System: The Eye
Interestingly, other animals have different arrangements of rods and cones in their
eyes. Many birds, for example, have a fovea in the middle of the retina for good central
vision but have a second, cone-rich fovea along the edge of the retina. Birds with this
visual arrangement include such diverse species as birds of prey, hummingbirds, and
swallows (Ruggeri et al., 2010). This second fovea gives these birds
accurate vision to their sides as well as in front of them, useful to
birds that must perform complex aerial movements. Other animals,
including many mammals such as dogs, have no foveae at all (Walls,
1963). The lack of a fovea is characteristic of animals that tend to be
nocturnal or rely primarily on senses other than vision. These animals
are more concerned about detecting moving objects than pinpointing
an exact location. Thus, rod vision is more critical than foveal vision.
In Figures 3.14 and 3.15, you can see a part of the retina called the
optic disc. The optic disc is the part of the retina where the optic nerve
leaves the eye and heads to the brain. In the optic disc, there are no
receptor cells. Axons of the retinal ganglion cells gather at the optic disc
and form the optic nerve, which carries the neural signal into the brain.
Because there are no receptors at the site of the optic disc, this location is
also called the blind spot. Because the blind spot in the left eye does not
correspond to the same region in space as the blind spot in the right eye,
humans really do not have a blind spot in their visual field. Even when
viewing the world with only one eye, we are not aware of our blind spot.
Higher areas of visual processing “fill in” the blind spot (Ramachandran,
1992). To observe your blind spot, see Figure 3.17. For an interactive
look at the blind spot, go to ISLE 3.5.
Retinal Physiology
Retinal anatomy concerns the structures of the retina. This refers to where cells are
located and how they connect with one another. Retinal physiology refers to the func-
tions of these structures—how they accomplish the processes that allow us to see. Because
we have reviewed the structures of the eye, we can now turn to how the eye converts
Optic disc: the part of the
retina where the optic nerve
leaves the eye and heads to the
brain; along the optic disc, there
are no receptor cells
200
150
100
50
10
Temporal
P
h
o
to
re
ce
p
to
r
d
e
n
si
ty
(t
h
o
u
sa
n
d
s/
m
m
2
)
NasalDistance from foveal center (mm)
0 10 20
20 0 20
Visual angle (degrees)
10 µm
O
p
ti
c
d
is
c
Rods
Cones
16.0 8.0 5.0 1.35 0.0 1.35 5.0 8.0 16.0
FIGURE 3.16 Receptor
Density in the Retina
The graph shows the density of
cones and rods in the retina. The
density of cones is highest in the
fovea, whereas cones become
very sparse as one moves away
from the fovea. Rods become
more common at the periphery.
Source: Curcio, C. A., Sloan, K. R., Kalina, R. E., & Hendrickson, A. E. (1990). Human photoreceptor topography.
Journal of Comparative Neurology, 292, 497–523.
FIGURE 3.17 Finding Your Blind Spot
Close your right eye and hold the book or your
laptop at arm’s length. Slowly move the book toward
you while looking at the orange circle with your left
eye. At some point, the plus sign will disappear. The
plus sign disappears when its image lands on the
optic disc, where there are no receptor cells.

66 Sensation and Perception
light into a neural signal at a physiological level. Retinal physiology is difficult, so read
carefully, review, and test yourself. Once you understand the physiology of transduction
in vision, the transduction processes for other senses will be more straightforward.
Transduction of Light
The rods and cones of the retina are equipped with the ability to convert light into a
neural signal. They accomplish this task by using chemicals known as photopigments.
Photopigments are molecules that absorb light and by doing so trigger events that alter
the voltage in the cell. When a photopigment absorbs a photon of light, it changes shape.
This change in shape initiates a series of biochemical processes, which result in a change in
electric potential, which allows a signal to exit the photoreceptor. In this way, a neural signal
leaves the photoreceptor and is, ultimately, transmitted to the optic nerve to be sent to the
brain. Once the photopigment has changed shape, it can no longer capture light. So other
biochemical processes reset the photopigment to its original shape so that it can detect more
light. Photopigments are contained in discs within the rods and cones. Each photoreceptor
may contain about 100 to 200 of these discs, and in each disc, there may be billions of pho-
topigment molecules (Fain, 2003). The many biochemical steps necessary to convert pho-
topigments from one form to another have been heavily investigated (see Rodieck, 1998).
Photopigments are composed of two molecules, which are bound together, a pro-
tein called an opsin in rods (and chromodopsin in cones) and a form of vitamin A
called retinal. The photopigment in the rods is called rhodopsin. The photopigments
in the cones are similar to rhodopsin, but there are three classes of these cone pho-
topigments (Bowmaker, 1998). The nature of the photopigment is central to deter-
mining how a receptor behaves with regard to light. These different photopigments
in each type of receptor tend to absorb different wavelengths of light, and it is the
differences in the opsins that are key to these differences. (See Figure 3.18 for a
Photopigment: a molecule
that absorbs light and by doing
so releases an electric potential
by altering the voltage in the
cell
Opsin: the protein portion of
a photopigment that captures
the photon of light and begins
the process of transduction; the
variation in opsin determines
the type of visual receptor
Retinal: a derivative of
vitamin A that is part of a
photopigment
FIGURE 3.18 Opsins
The figure shows a depiction of the four most common opsins found in human eyes. Panels (a) and (b) show the rod opsin. Panel (c) at the bottom
shows the three opsins found in the different classes of cones. The different color dots show how each opsin is different from the opsin to its left.
Source: Nathans et al., 1986.
(a)
(b)
(c)

67 Chapter 3: Visual System: The Eye
depiction of the rod and the
three cone options and how
they differ from each other.)
The role of retinal in the rho-
dopsin explains one piece
of folk knowledge you may
have heard about. You may
have been told that eating
carrots will help your night
vision. Carrots are very rich
in vitamin A, which is the
source of retinal. If you have
a healthy diet, you will proba-
bly not improve your vision by
eating more carrots, but diets
that are deficient in vitamin
A can result in night blind-
ness, that is, an inability to
see under dim light conditions
(Wald, 1968).
In the photopigment that
is ready to respond to light,
retinal is in a bent form, called
11-cis, that fits into the opsin.
When a photon of light is
absorbed by the photopigment, the rhodopsin straightens out and the photopigment
breaks apart (Wang, Schoenlein, Peteanu, Mathies, & Shank, 1994). This straight-
ening of the bond takes place during an incredibly fleeting time period (Schoenlein,
Peteanu, Mathies, & Shank, 1991). See the two shapes of retinal in Figure 3.19. The
effect of this absorption of the photon by the photopigment causes the receptor to
have a more negative voltage inside relative to outside the receptor (Tomita, 1970).
This is called hyperpolarization, referring to the polarization of electric charge
increase. This hyperpolarization causes the receptor to release less of the neurotrans-
mitter, which are molecules that neural cells release to communicate with subsequent
neural cells. In this way, the receptor can communicate with the rest of the retina.
Thus, oddly, light is inhibitory to receptors. So, the question naturally arises, how
does light act to stimulate the visual system so that we can see? It turns out that the
neurotransmitter of the rods, probably glutamate, is inhibitory to the other cells of
the retina; that is, it stops the responses of these bipolar and horizontal cells (Schiells
& Falk, 1987). Thus, light inhibits the release of an inhibitory neurotransmitter and,
ultimately, excites the visual system, like multiplying two negative numbers and
getting a positive number. There is no good explanation for why the visual system
evolved this seemingly illogical system.
Given that one photon of light can change a photopigment molecule in one
receptor and induce a neural signal, a natural question to ask is how many pho-
topigment molecules must be activated to create a perception of light, to reach
absolute threshold? In particular, because we know that rods are sensitive to small
amounts of light, we will look to the photoreceptors at the periphery to answer
this question. In fact, if you optimize the conditions for light detection, a single rod
may start a signal in response to as little as one photon of light, and a person may
actually see a light with as few as seven photopigment molecules activated (Hecht,
Schlaer, & Pirenne, 1942).
Hyperpolarization: a change
in the voltage of a neuron
whereby the inside of the cell
becomes more negative than it
is in its resting state
Neurotransmitter: a chemical
substance neural cells release
to communicate with other
neural cells
FIGURE 3.19 Transduction at the Molecular Level
The photopigment molecule has one of two shapes, each called an isomer of the other. In
the photopigment that is ready to respond to light, retinal is in a bent form, called 11-cis, that
fits into the opsin. When a photon of light is absorbed by the photopigment, the rhodopsin
straightens out and the photopigment breaks apart. This straightening of the bond takes place
during a very short time period. The effect of this absorption of the photon by the photopigment
causes the receptor to have a more negative voltage inside relative to outside the receptor.
H
3
C
H
3
C
HC
C
C
C
C
C
C
C
11-CIS RETINAL
ALL-TRANS RETINAL
C
C
C
H
3
C
H
2
C
H
2
C
H
2
C
H
2
C
CH
3
CH
3
CH
3
CH
3
CH
3
CH
3
CH
3
C
H
2
C
H
2
CH
CHC
H
C
H
C
H
C
H
C
H
C
H
H
C
H
C
H
C
H
C
H
C
O
O
Photoisomerization
Photon of light
Photopigment
regeneration

68 Sensation and Perception
Classes of Receptors
The human eye uses four different photopigments, one for the rods and three in differ-
ent cone classes in the retina (Figure 3.20). Each of these photopigments is maximally
sensitive to a particular frequency of light (Table 3.2). It is the interaction of these three
photopigments and subsequent perceptual processing that provides for our complex
color vision. This classification system is actually a simplification, as each class of cone
photopigment has many subtypes, especially in the cones that respond to red and green.
In rare cases, some people (most often women) have a fourth cone class with a fourth
photopigment. There is some evidence that these people see a greater range of colors
than do normal people (Jameson, Highnote, & Wasserman, 2001).
TEST YOUR KNOWLEDGE
1. Summarize the mechanism of transduction in the rods and cones.
2. Determine the reason we have more rods than cones in our eyes and predict what impact
having the cones in one area on our retina will have on how we use our visual system and
eyes.
THE DUPLEX THEORY OF VISION
3.4
Illustrate how the visual system adapts to the dark and then adapts
again to the light to help us adjust to changing light levels.
The duplex theory of vision is the idea that there are functionally two distinct ways in
which our eyes work. The first system is called the photopic system, and the second is
the scotopic system. The photopic system is associated with the cones, and the other,
the scotopic system, is associated with the
rods. The duplex theory is the idea that our
visual system can operate in fundamentally
different ways, depending on the conditions
in the environment. Daytime is the domain
of photopic vision, but nighttime is when
our scotopic vision comes to the forefront.
These hypothesized differences in visual
function derive from the following observa-
tions. First, rods are more sensitive to light
overall than are cones. Second, rods are
most sensitive to different wavelengths than
cones are. That is, they have a different spec-
tral sensitivity. Third, rods and cones have
different spatial and temporal summation
properties, which we discuss shortly. Finally,
cones support color vision, but rods do not.
In general, the photopic system is associated
with daytime vision, and the scotopic sys-
tem is our night vision system, but there is a
range of intermediate ambient light intensity
within which both systems work. This inter-
mediate zone is said to be mesopic vision.
Duplex theory of vision:
the doctrine that there are
functionally two distinct ways
in which our eyes work, the
photopic, associated with
the cones, and the scotopic,
associated with the rods
Photopic vision: the vision
associated with the cones; it is
used in the daytime, has good
acuity in the fovea, and has
color vision
Scotopic vision: the operation
of the visual system associated
with the rods; it has relatively
poor acuity and no color ability
but is very sensitive to light
FIGURE 3.20 Wavelength Sensitivity of Cones and Rods
The photopigments in each cone and in the rods vary with respect to their ability
to absorb light at different wavelengths. Rods show a peak sensitivity at just shy
of 500 nm, which would be seen as dark green by the color system. With respect
to the cones, the S-cones’ peak sensitivity is at 420 nm, the M-cones’ at 530 nm,
and the L-cones’ at 560 nm. Note that each photoreceptor responds to a broad
band of wavelengths but at less strength than to the peak. We determine color by
a complex comparison of the responses from each of the cone photoreceptors.
100
50
R
e
la
ti
v
e
s
e
n
si
ti
v
it
y
(
%
)
0
400 450 500 550
530
Wavelength (nm)
560420380
S
-c
o
n
e
s
Ro
ds
M
-cones
L-cones
600 650 700

69 Chapter 3: Visual System: The Eye
See ISLE 3.6 to see differences between photopic vision and scotopic vision. Table 3.3
also illustrates the differences between these two parts of the duplex system.
Spectral Sensitivity and the Purkinje Shift
Spectral sensitivity refers to the relative sensitivity of a receptor type to all of the
wavelengths. Rods are relatively more sensitive to the shorter wavelengths, whereas
cones are relatively more sensitive to the longer wavelengths. Indeed, cones, taken as a
whole, are most sensitive to about 555 nm, and rods are most sen-
sitive to just under 500 nm. This difference in spectral sensitivity is
named the Purkinje shift, after the Czech anatomist Jan Evangelista
Purkyně. He would take long walks during dawn and noticed that
his favorite bright red flowers appeared very dark before the sun
climbed high into the sky (Wade & Brožek, 2001). This shift occurs
as we transition from day vision to night vision and back again. We
have simulated these differences in the relative sensitivity of rods
and cones to long- and short-wavelength light in ISLE 3.7.
An interesting practical implication of the Purkinje effect is
that, as nighttime conditions appear, longer wavelengths of light
will appear darker, whereas shorter wavelengths will appear rela-
tively brighter. This means that red objects become more difficult
to see at night than blue or green objects with similar reflectance
values. It is for this reason that, some years ago, fire departments
started shifting away from traditionally red vehicles. Because
red is harder to see at night, fast-moving red fire trucks will be
more difficult for other drivers to see than fast-moving blue fire
trucks. Nonetheless, tradition is often hard to break, and many
fire departments still opt for the traditional fire-engine red, even
though it puts both firefighters and other drivers at more risk. However, another
driving convention does have perceptual validity. Red light is less likely to affect our
ability to adjust our eyes to see in the dark (dark adaptation) than other colors. This
validates the decision to make taillights on cars red. Drivers will see red taillights
at night, but the red light will not affect their dark adaptation as much as taillights
of other colors would. You may have noticed how distracting a white (or broken)
Purkinje shift: the observation
that short wavelengths tend to
be relatively brighter than long
wavelengths in scotopic vision
versus photopic vision
ISLE 3.6
Photopic vs. Scotopic Vision
ISLE 3.7
Purkinje Shift
Characteristic Photopic Scotopic
Color vision Yes (three cone
classes)
No (only one type of rod)
Peak spectral sensitivity ~550 nm ~505 nm
Acuity (spatial summation) Good Poor
Temporal resolution (temporal
summation)
Good Relatively poor
Regions of greatest sensitivity Fovea ~20 degrees at the
periphery
Relative sensitivity 1,000,000 1
Receptor
Type
Peak Sensitivity
(Nanometers)
Rods 496
S-cones 420
M-cones 530
L-cones 560
TABLE 3.2 Receptor Types
NOTE: Each receptor type shows a peak sensitivity to light at a
particular wavelength. The receptor type is shown on the left,
and the peak sensitivity (in nanometers) is shown on the right.
These peak sensitivities are approximate and may vary among
people with normal vision.
TABLE 3.3 Photopic Versus Scotopic Vision

70 Sensation and Perception
taillight can be at night, as it interferes with your ability to see well in the dark. This
is also why the little screens of cell phones can be so distracting in an otherwise dark
movie theater.
Spatial Summation and Acuity
The next difference between photopic and scotopic vision is the relative acuity of the
two systems. Acuity refers to the ability to see or resolve fine details. The reason for
the acuity difference is partially that rods and cones have different spatial summation
characteristics. Spatial summation refers to the ability to pool light across different
regions of space. At night, you want to detect as much light as possible, even if spatial
summation causes things to be a bit blurry. During the day, there is ample light, so
acuity becomes the prominent issue. As you know by now, rods are more sensitive to
dim light than cones are. Part of this light sensitivity occurs because many rods connect
to one retinal ganglion cell where their responses are added together. This allows the
scotopic system to pool responses across different rods in order to maximize sensitivity
to light. However, by pooling across rods, the scotopic system loses some ability to
resolve light to particular locations in the visual world. This pooling of information is
called convergence, and it is much greater for rods than for cones. The lowest level
of convergences is for cones in the fovea, thus maximizing the ability to pinpoint the
source of the light in space, even if it does not allow sensitivity to dim lights. In fact,
in the fovea our acuity is determined by the density of cones, whereas in the periphery,
even in photopic vision, our acuity is determined by the density of a particular type of
ganglion cell, the midget cell, that we will discuss later (Rossi & Roorda, 2010). See
ISLE 3.8 to explore how changes in convergence alter how we see details. Try the image
provided or upload one of your own.
To illustrate convergence with an example, consider admiring the fine coat of a
Bengal cat (Figure 3.21). You want to be able to make out the intricate patterns on
the coat and see them in fine detail. To do this, you cannot confuse one location along
the cat’s fur with another. Thus, neutrally, you want as little spatial summation as
possible. This is what the cones provide. The cones, particularly in the fovea, pick
up light in one very small area and project that through one ganglion cell, so as not
to cause spatial confusion. Thus, cones allow us to see with greater visual acuity,
whereas rods are more sensitive to low levels of light.
Dark and Light Adaptation
Imagine waking up in the middle of the night. It is dark, and there are other people still
asleep in your house, so you do not want to turn the lights on and alarm them. But you
are hungry, and you want a snack. You know that there are still some of your favorite
brownies in the kitchen. You feel your way along the wall into the kitchen and enjoy
the chocolate brownies in the middle of the night. With your late-night craving satisfied,
you are ready to return to sleep. But before you return to bed, you may notice that after
you have been up for a little while, you can see in the dark. What before was obscured
you can now discern. This is the process of dark adaptation. It takes nearly 30 minutes
for the process to transition from seeing in bright day to seeing your best in the dark,
but the shift from photopic to scotopic vision allows people to see much better in the
dark. See ISLE 3.9a and 3.9b to see how fast we can dark adapt and a simulation of our
changing sensitivities on dark adaptation.
What is going on during dark adaptation? First, a definition: Dark adaptation
is the process in the visual system whereby its sensitivity to low light levels is
increased. During the first 8 minutes of trying to see in the dark, it is still the
cones that are doing most of your seeing. Then, your sensitivity to low levels of
light increases, as your scotopic system rods come online and start taking over.
ISLE 3.8
Convergence
Convergence: the number of
photoreceptors that connect
to each ganglion cell; more
convergence occurs for rods
than for cones
Dark adaptation: the process
in the visual system whereby its
sensitivity to low light levels is
increased
ISLE 3.9
Dark Adaptation Function

71 Chapter 3: Visual System: The Eye
Part of the explanation for dark adaptation includes the properties of the pho-
topigments themselves. Dark-adapted eyes have a higher concentration of rho-
dopsin, the active pigment in the rods. In the darkness, the levels of rhodopsin
increase. Thus, one factor in dark adaptation is the concentration of photopig-
ment (Rushton, 1965). Other factors include dilation of the pupil, which happens
rapidly in lower light levels, and other factors in how the retina processes light
(Figure 3.22 and ISLE 3.9).
Light adaptation occurs when light returns, such as when electric lights are turned
on in an otherwise dark room. It is acceptable to think of light adaptation as being
like dark adaptation but in reverse. However, there is one major difference: Light
adaptation will be completed in about 5 minutes, or even less, as the visual system
switches back to relying on cones rather than rods. Dark adaptation is therefore
much slower. It seems that this difference is due partly to the fact that light adap-
tation is driven actively by the light entering the eye, whereas dark adaptation is a
more passive response to the lack of light. Of course, under natural conditions, in
which the only source of light is the sun, dark adaptation is sufficiently fast to adjust
to changing light conditions as the sun sets, and light adaptation might be mostly
gradual as well. The change from afternoon sun, through dusk, to evening is about
the time it takes dark adaptation to complete its total swing.
There are some interesting implications of the shift from light adaptation to dark
adaptation and back again, especially in our constructed world, in which a person
may go back and forth from lit rooms to dark ones every
few minutes. Think about a police officer, patrolling at
night. He should be dark adapted to survey the nighttime
scene, but may need to shift to photopic vision to read the
computer in the patrol car. If a task requires a person (such
as our police officer) to go quickly from a lighted area to a
dark area in a short period of time, then the period required
for full dark adaptation may present problems. If the offi-
cer, after checking his computer, must enter a dark alley
in pursuit of a suspect, his vision may not be optimized.
This dark adaptation period delays how fast the person can
begin functioning in the dark. However, this limitation can
be overcome by having the person work in an environment
with only red lights. The red light will not be absorbed by
the rods as well as other wavelengths, allowing the dark
adaptation process to begin in the rods while the cones
respond to the red light. Then, when the light is removed,
the person is dark adapted to a large degree and can oper-
ate immediately in the dark, an advantage for a police offi-
cer in the field. This works because the cones are relatively
more sensitive to long-wavelength light than the rods. The
greater sensitivity of cones than rods to the long wavelengths is so significant that
cones are actually more sensitive to these wavelengths than rods are in an absolute
sense, not just a relative sense (Cornsweet, 1970).
A common example of this technique is the use of red flashlights when observing
with a telescope at night. An observer can look at a star chart or other information
using the red flashlight. This will give her the ability to use her photopic vision to read
the details of the star chart. Then, when the light is turned off, she still has her scotopic
rod vision to see the subtle detail in the night sky, such as a faint galaxy or nebula. In
modern smartphones, apps are available to help you determine which star is which
when you stargaze. Smartphone apps, such as Google Sky Map, can be put into night
mode, in which the display is in red.
FIGURE 3.21 The Fine
Coat of a Bengal Cat
Our photopic system allows us
to see the fine detail of patterns
and colors in the coats of these
beautiful cats.
©
iStockphoto.com
/photodeti
Dark adaptation
curve for rods only
6
5
4
3
2
1
0
0 5 10 15
Minutes in dark
B
ri
g
h
tn
e
ss
o
f
li
g
h
t
th
a
t
is
j
u
st
b
a
re
ly
v
is
ib
le
20 25 30
Dark adaptation
curve for cones only
Maximum cone
sensitivity
Maximum visual
sensitivity
Overall dark adaptation
curve (for rods and
cones combined)
Rod-cone
break
FIGURE 3.22 The Process of Dark Adaptation
As a dark period extends in time, sensitivity to light increases.
In the graph, the blue line shows the overall increase in
sensitivity to light. The orange line shows the change in
sensitivity for cones in the fovea.
Light adaptation: the process
whereby the visual system’s
sensitivity is reduced so that
it can operate in higher light
levels

72 Sensation and Perception
Another application of the knowledge of dark adaptation takes place in the
modern aircraft cockpit. There are a lot of visual situations that can alter a person’s
ability to read a cockpit display. Think of a pilot flying with the sun directly in his
face. The pilot may light-adapt to the sun and then not have sufficient sensitivity
to read the electronic display when needed. The opposite problem occurs at night,
when it is desirable to not have the displays so bright that the pilot will light-adapt
to the display and not be able to see the runway clearly. To avoid these problems,
researchers determined the light levels needed to see the sky in front of the airplane
and also to see the displays inside the cockpit in situations with high light intensity
and low light intensity. From these data, Krantz, Silverstein, and Yeh (1992) devel-
oped automatic adjustments so that pilots do not have to spend time adjusting the
luminance of their displays when the lighting conditions change. When it is dark
outside, the displays inside the cockpit will switch to a red light that allows the
pilots to dark-adapt for seeing outside the plane but also allows them to read their
instruments inside the plane (Figure 3.23). Some smartphones have incorporated
some of this research. See ISLE 3.10 for how automatic lighting adjustment works
in airplanes.
Development: Infant Acuity
Parents and grandparents stare intently at their newborn child or grandchild, but what
does the child see if they are looking back (Figure 3.24)? Studies of the developing eye
indicate that although the rods are nearly adultlike in their development by birth, the
cones are not. Cones are not packed as closely together in the fovea of children as they
are in the fovea of adults (Banks & Bennett, 1988). As discussed earlier, the density
of cones in the fovea is critical to the high level of acuity we enjoy as adults (Rossi &
Roorda, 2010). If infants have a lower density of cones at birth, then they should have
a lower acuity, but how do we determine what their acuity is?
In Chapter 2, we discussed psychophysics, which depends on asking the partic-
ipant a question such as “Did you see the stimulus?” A researcher could try that
with an infant, but it is unlikely to be very successful. Researchers must consider
the fact that infants do not talk, but they get fussy easily and are not noted for
their patience with researchers’ interests. As such, researchers have been very cre-
ative in developing methods that allow us to examine the visual abilities of infants.
Some methods rely on the observation that infants show preferences for some
stimuli over others by various behaviors, such as widening of the eyes or, when a
little older, looking in one direction or another. Other methods employ variations
of recording brain activity such as electroencephalography
(EEG) and event-related potentials (ERPs). In all of these
methods, two stimuli are presented, one an acuity stim-
ulus (such as alternating white and black bars) and the
other a plain gray field. If the infant responds differently
to the two stimuli, preferring the bars to the gray field, or
having a different brain activity, then it is determined that
the infant can see those bars. Thus, we can determine an
infant’s acuity by finding the smallest bars that the infant
responds to.
Several studies have examined infant acuity with a variety of
these methods (e.g., Atkinson, Braddick, & Moar, 1977; Mayer &
Dobson, 1982; Salomão, Ejzenbaum, Berezovsky, Sacai, &
Pereira, 2008; Sokol, 1978), and although the absolute values
of acuity often disagree, they all find the same general pattern as FIGURE 3.23 Dark Adaptation in an Aircraft Cockpit
®
R
og
er
R
es
sm
ey
er
/C
or
bi
s/
VC
G
/C
or
bi
s
D
oc
um
en
ta
ry
/G
et
ty
Im
ag
es
ISLE 3.10
Automatic Light
Adjustment in Cockpits

73 Chapter 3: Visual System: The Eye
shown in Figure 3.25. An infant’s acuity is quite poor when he is
born but it develops rapidly, particularly in the first year of life.
However, the child does not have adult acuity (indicated by the
Snellen acuity of 20/20 on the figure) at that time. Acuity con-
tinues to improve over the next few years but at a much slower
rate (Figure 3.25). There is a reason that when children are first
learning to read, the letters are larger than in books for adults.
TEST YOUR KNOWLEDGE
1. Diagram the different elements of the visual system that allow
us to adjust to a wide range of lighting conditions.
2. Devise an outline that describes the difference between
daylight vision and nighttime vision.
RETINAL GANGLION CELLS
AND RECEPTIVE FIELDS
3.5
Diagram the receptive fields of retinal ganglion cells,
and show how they need contrast to operate.
Each photoreceptor is connected to several different retinal
ganglion cells, which convey the signal into the optic nerve and
toward the brain. There are only about 1.1 million ganglion
cells, many fewer than the total number of rods and cones.
Therefore, most ganglion cells receive inputs from many dif-
ferent photoreceptors (called convergence; see Figure 3.26).
A ganglion cell receives input from receptors that are next to
each other on the retina, which form a grouping of adjacent
receptors. This grouping is called a receptive field. A receptive
field is the area in the visual world that a particular vision neu-
ron responds to. But in this case, it also refers to the array of
photoreceptors from which each retinal ganglion cell receives
input. In the retina, the first visual receptive fields studied were
the receptive fields of the ganglion cells. Kuffler (1953) was the
first researcher to study these remarkable cells. A particular
ganglion cell that receives input from the fovea might have a
receptive field that corresponds to exactly the point you are
looking at. A ganglion cell that receives input from the far
periphery might have a receptive field that corresponds to a
point in space just at the edge of your visual field, away from your point of central fix-
ation. Receptive fields are illustrated in Figure 3.26. Examine this figure carefully; there
is a lot of information in it.
In the fovea, just a few cones map onto one retinal ganglion cell. Therefore, the
receptive fields of these ganglion cells are quite small and correspond to small regions
in visual space, consistent with the idea that the fovea is designed for high visual acu-
ity. In the periphery, however, many cones or rods, in some cases hundreds of them,
will map onto a single ganglion cell. As you move away from the fovea, ganglion
receptive fields get progressively larger (Watanabe & Rodieck, 1989). These greater
receptive fields allow greater convergence and greater sensitivity to dim light. This
FIGURE 3.24 Parents With Newborn Baby
Parents and grandparents enjoy staring at newborns, but
what does the newborn see?
20/20
20/32
20/50
20/80
20/125
20/200
20/320
0 6 12 18 24 30 36
S
n
e
ll
e
n
a
cu
it
y
Age (months)
Mean
FIGURE 3.25
Development of Visual Acuity in Young Children
Children show a rapid development in acuity over the first
year of life and then a slower development of acuity to adult
levels in the next few years.
Source: From Salomão, S. R., Ejzenbaum, F., Berezovsky, A., Sacai, P. Y., &
Pereira, J. M. (2008). Age norms for monocular grating acuity measured
by sweep-VEP in the first three years of age. Arquivos Brasileiros de
Oftalmologia, 71(4), 475–479.
Receptive field: a region of
adjacent receptors that will
alter the firing rate of a cell
that is higher up in the sensory
system; the term can also apply
to the region of space in the
world to which a particular
neuron responds
Jennifer Volker Krantz

74 Sensation and Perception
summation across adjacent rods is called convergence (reviewed earlier in the section
on spatial summation and different from the meaning of convergence as applied to
the lens). Thus, retinal ganglion cells are processing for acuity in the fovea and light
detection in the periphery.
Retinal ganglion cells also start the process of edge detection. Critical to vision,
edge detection involves determining the location at which one object ends and the
next object begins. This is important in discriminating objects, detecting camouflage,
and many other important visual functions. We focus on edge detection in terms of
perception later. But for now, it is important to know that the visual system searches
for these edges and enhances them, so that we can clearly see them. The earliest stage
of edge detection occurs in the retinal ganglion cells and is known as center-surround
receptive fields (Figure 3.27 and ISLE 3.11).
Kuffler (1953) noticed during single-cell recording of retinal ganglion cells that he
could stimulate the center of a receptive field, resulting in a different response from
the cell than when he stimulated the surrounding area of the receptive field of that
cell. In some cases, stimulating the center of the field resulted in an increase in the
number of neural responses, called excitation. In other cells, stimulating this same
part of the receptive field resulted in a decrease in the number of neural responses,
called inhibition. Note that both a positive (excitation) and a negative (inhibition)
High
Moderate
Firing rates
Low
Baseline
High
convergence
High convergence, low acuity
Same signal sent
into optic nerve
Different patterns of signals
sent into optic nerve
No convergence, high acuity
Moderate
convergence
No
convergence
Cones
Incoming light
Neural
signals
Retinal
ganglion
cell
Action
potentials
(a) Degrees of convergence
Low firing
rate
Moderate
firing rate
High
firing rate
(b) Spatial summation
(c) Acuity
FIGURE 3.26 Convergence of Photoreceptors Onto Ganglion Cells
In the fovea, there is little convergence or no convergence, as one receptor cell matches onto one ganglion cell. However, as one
moves into the periphery, there is more convergence, as multiple photoreceptors match onto the same ganglion cell. The more
convergence, the greater the light sensitivity, whereas the less convergence, the greater the acuity.
Edge detection: the process
of distinguishing where one
object ends and the next
begins, making edges as clear
as possible
Center-surround receptive
field: a receptive field in which
the center of the receptive field
responds opposite to how the
surround of the receptive field
responds; if the center responds
with an increase of activity to
light in its area, the surround
responds with a decrease in
activity to light in its area

75 Chapter 3: Visual System: The Eye
response are still responses, just of different kinds. If you stimulate an area outside a
particular cell’s receptive field, you will get no response, neither excitation nor inhi-
bition. This difference in excitation and inhibition allows the cell to start coding for
edges. We will explain why shortly.
There are two kinds of center-surround receptive fields. They are known as
on-center receptive fields and off-center receptive fields. In on-center receptive
fields, the cell’s center produces excitation, whereas the surround produces inhibi-
tion. In off-center receptive fields, the center produces inhibition, whereas the sur-
round produces excitation. In either case, if you present light that covers the entire
receptive field of the cell, the response will be small, as the excitation and inhibition
will cancel each other out. However, if you present a pattern with edges (contrasting
light and dark), the cell may respond strongly. That is, these cells respond to lumi-
nance contrast, or differences in light intensity across the field. To make sense of this,
think of the following: A uniform field in the visual world is usually not interesting
or important (think of a blank wall or a uniformly blue sky). Contrast, however, is
(a) ON-center ganglion cell
(b) OFF-center ganglion cell
Spot in
center
Spot in
surround
Light on
Light on
Response
Spot in
center
Response
Response
+
+
+ + −



+
+
+ + −



+
+
+ + −



Spot in
surround
Response
+
+
+ + −



FIGURE 3.27
Center-Surround Receptive Fields
(a) An on-center ganglion cell. The image shows
a receptive field for an on-center ganglion cell.
When a light is shined in the center, the ganglion
cell becomes more active than its base-rate
state. However, when a light is shined at the
periphery of the receptive field, the ganglion cell
is inhibited below base rate. (b) An off-center
ganglion cell. The image shows a receptive field
for an off-center ganglion cell. When a light is
shined in the center, the ganglion cell becomes
less active or inhibited relative to its base-rate
state. However, when a light is shined at the
periphery of the receptive field, the ganglion cell
becomes more active than base rate.
ISLE 3.11
Center-Surround Receptive Fields
On-center receptive fields:
retinal ganglion cells that
increase their firing rate
(excitation) when light is
presented in the middle of the
receptive field and decrease
their firing rate (inhibition) when
light is presented in the outside
or surround of the receptive field
Off-center receptive fields:
retinal ganglion cells that
decrease their firing rate
(inhibition) when light is presented
in the middle of the receptive
field and increase their firing
rate (excitation) when light is
presented in the outside or
surround of the receptive field

76 Sensation and Perception
important—whether it be the change of lightness where the wall meets the door or
the roiling patterns of clouds of an impending thunderstorm. Or think about our
police officer on patrol at night. She wants to be able to make out the dim form of a
prowler as distinguished from the background night. Thus, the visual system should
be looking for these contrasts, as they indicate interesting features of the visual world.
Indeed, it appears that edge detection, such as what is seen in center-surround orga-
nization, starts early in visual processing, already occurring in the retinae themselves.
For an illustration of this, go to ISLE 3.11 to simulate how Kuffler (1953) conducted
his experiments and see how these receptive fields need contrast to respond.
The physiological mechanism that creates center-surround receptive fields is quite
complex. To understand how it works, one must consider the roles of horizontal cells
and bipolar cells, as well as the rods, cones, and retinal ganglion cells. Consider the
receptive field of retinal ganglion cells in the foveal region of the retina. A person
directs his or her gaze to an object, such as a cup of coffee on a table. The image of the
coffee cup is now projecting directly onto the fovea. Cones in the fovea will initiate an
excitatory signal in response to the light reflecting off the coffee cup. This signal will
move to both the bipolar cells and the horizontal cells. In response to the signal from
the cones, horizontal cells send inhibitory signals back to adjacent cones. Horizontal
cells also send excitatory signals across to other horizontal cells, thus increasing the
inhibitory feedback on the cones. That is, these cells act to minimize the responses to
uniform fields (an empty table) and maximize the response to changing fields (a coffee
cup on that table). However, this inhibition decreases across space, so that cones far-
ther from the source of the image are less inhibited. The combination of the positive
signals coming from bipolar cells and the inhibition coming from the horizontal cells
determines the strength of a signal to a retinal ganglion cell, which then has its own
center-surround characteristics. The result of these complex processes is known as lat-
eral inhibition (Hartline & Ratliff, 1958).
The goal of lateral inhibition is to facilitate edge detection. Edge detection is the
result of lateral inhibition which is the outcome of center-surround receptive fields.
That is, it is lateral inhibition working in receptive fields that make edges as clear as
possible. Edge detection is important—think about trying to see a camouflaged pred-
ator, perhaps a lioness, in the tall dry grass. The lioness’s fur and the grass are roughly
the same color, important for the lioness as she slowly stalks her prey (you, in this case).
However, the fur and the grass are not identical, and you want your visual system to
pick up where the grass ends and where the lion begins (so you can jump back into
your Land Rover and speed away and have stories to tell for a long time). This involves
edge detection and gestalt figure–ground discrimination (see Chapter 5). Some animals
have developed patterns on their fur or skin to confuse edge detection. For example,
the stripes on tigers break up the outlines of their body and make them harder to spot
when partially hidden. But if we think of it in terms of figuring out where one object
ends and the next begins, we have the basic function of edge detection (Figure 3.28). It
is for this reason that edge detection evolved. As we will see throughout our discussion
of the visual system, edge detection is built in at every level.
Lateral inhibition explains a famous visual illusion known as Mach bands, named
after their discoverer, Ernst Mach (1838–1916). Mach bands are illustrated in
Figure 3.29. Examining the pattern in the figure, you see bands of dark gray on the
left and progressively lighter bands of gray to the right. If you look at each band, it
looks as if it is slightly lighter on the left side than on the right side. This is an illu-
sion. The reflectance of each band is uniform. That is, the perceived intensity appears
to take a big step at the border of each band and then gradually declines until you get
to the next edge. This is edge enhancement. Lateral inhibition accentuates the edges
of the stimulus, so that you see the borders of reflectance more clearly. However, it is
also an illusion, as each band has the same reflectance across its space. In the Mach
Lateral inhibition: the
reduction of a response of
the eye to light stimulating
one receptor by stimulation of
nearby receptors, caused by
inhibitory signals in horizontal
cells

77 Chapter 3: Visual System: The Eye
bands, the enhanced edges are illusory. But the system has functional value—in
the real world, perceiving shape is critical, and lateral inhibition helps us do so. See
ISLE 3.12 for an interactive version of Mach bands.
TEST YOUR KNOWLEDGE
1. Create a diagram for yourself to show the spatial summation of rods and cones to create
a center-surround receptive field.
2. Critique the operation of center-surround receptive fields. How do they enhance vision
and create illusions?
REFRACTIVE ERRORS AND
DISEASES OF THE EYE
3.6
Judge the different refractive errors and eye diseases
on how they affect the ability of the eye to see.
As you can tell from reading this chapter, the human eye is an extremely complicated
organ, and vision is a very complex process. The result, seeing, is remarkable when
you consider all the eye and brain must do in such a short time to allow us to see
FIGURE 3.28
Edge Enhancement
These Thomson’s gazelles are alert
for the possibility of approaching
predators. (a) A normal photograph.
(b) Same photo with edges
enhanced. This kind of enhancement
allows better detection of objects
against a uniform background.
©
S
hu
tt
er
st
oc
k.
co
m
/
Th
eL
ea
rn
in
gP
ho
to
gr
ap
he
r
Isle 3.12
Mach Bands
(a) (b)
Stimulus
(b)
Receptors
Ganglion cells
What you
experience
1 2
2
3
3
4
4
5
5
6
6
7
7
8 9 10 11 12 13 14 15
8 9 10 11 12 13 14 151
FIGURE 3.29 Mach Bands
(a) Each bar in this picture is uniform in lightness. However, we see a different level of lightness in each bar from left to right. For each bar, the
left edge looks blacker than the right edge, which looks lighter or whiter. (b) This illusion is the result of processes of edge detection, such as
lateral inhibition. When cells with receptive fields that straddle the bars detect the lightness contrast, they act to enhance the edges.
(a)

78 Sensation and Perception
as we do. Nonetheless, problems can arise that interfere with clear visual perception.
Undoubtedly, you know people who wear eyeglasses or contact lenses, and you may
wear them yourself. The origin of these minor problems is interesting and is discussed
in this section. You may also know someone with more serious visual problems, such as
cataracts or macular degeneration. We discuss these two disorders as well.
Myopia (Nearsightedness)
Myopia is a common form of mild visual impairment, often called nearsightedness
because people with myopia can focus well on near objects, but faraway objects appear
blurry. There is usually a parallel problem with both eyes,
but a person can potentially be myopic in one eye but not
the other. Moreover, if the person has myopia, it is typ-
ical that each eye has a different level of myopia. With
myopia, the eye tends to be too long from front to back
for the lens. The process of accommodation cannot make
the lens thin enough to focus the light from more dis-
tant objects onto the retina. Because these distant objects
are focused in front of the retina instead of onto it, the
objects appear blurry. Another way to state this is that
the lens is too strong for the length of the eye. To correct
the problem, the lens must be weakened. To weaken the
lens, a diverging or negative artificial lens is used, that is,
typical eyeglasses. A diverging or concave lens is wider
at the edges than at the middle; this is the opposite of
the eye’s own lens and the lenses that have been used in
the figures. As a result, light spreads out even more after
it has passed through the lens than it would otherwise.
Eyeglasses or contact lenses refract or bend light just
enough so that the image can be focused onto a person’s
retinae, allowing a focused image. However, once the
glasses or contacts are removed, light comes into focus
in front of the lens, and objects at a distance will again
appear blurry, though objects up close will be in focus.
Thus, myopia is readily compensated for with glasses (or
contacts), but people must wear their eyeglasses in order
to compensate for this condition. The optics of the lens
in myopia are illustrated in Figure 3.30. For a dynamic
demonstration, see ISLE 3.13. Myopia is quite common. There are an estimated
70 million people in the United States alone with myopia.
Hyperopia (Farsightedness) and
Presbyopia (Old-Sightedness)
Both hyperopia and presbyopia are common forms of mild visual impairment, often
called farsightedness because people with these impairments can focus well on faraway
objects, but near objects may appear blurry. With hyperopia, the eye tends to be too
short for the lens (Figure 3.31). The lens is not strong enough for this eye. The process
of accommodation does not make the lens thick enough to focus the light from closer
objects onto the retinae. Because these close objects are focused behind the retinae
instead of on them, the objects will appear blurry. Distant objects, however, are seen just
fine. As with myopia, eyeglasses can be fitted that will allow near objects to be imaged
Myopia: a condition causing
an inability to focus clearly
on far objects, also called
nearsightedness; occurs
because accommodation
cannot make the lens thin
enough
Focus on retina
Light from
distant object
Normal vision
Normal optic axis
Focus in front
of retina
Thin lens
Thin lens
Light from
distant object
Myopia (Nearsightedness)
Long optic axis
FIGURE 3.30 Myopia
If the eye is too long internally, the lens cannot become thin
enough to focus an image on the retina. Instead, the image
focuses in front of the retina, making the image on the retina
blurry. Myopia is correctable with eyeglasses.

79 Chapter 3: Visual System: The Eye
ISLE 3.13
Correcting Myopia and Hyperopia
Hyperopia: a condition
causing an inability to focus
on near objects, also called
farsightedness; occurs because
accommodation cannot make
the lens thick enough
Astigmatism: a condition
that develops from an irregular
shape of the cornea or the lens,
which makes it impossible for
the lens to accommodate a fully
focused image
FIGURE 3.31 Hyperopia, or Farsightedness
If the eye is too short internally, the lens cannot become thick
enough to focus an image on the retina. Instead, the image
focuses behind the retina, making the image on the retina blurry.
Hyperopia is correctable with eyeglasses. Similar to hyperopia
is presbyopia, in which a hardened lens tends to focus images
behind the retina. Presbyopia is very common as people age into
their 40s and 50s.
Focus on retina
Normal vision
Normal optic axis
Focus behind
retina
Thick lens
Thick lens
Light from
nearby object
Light from
nearby object
Hyperopia (Farsightedness)
Short optic axis
on the retinae. These glasses use converging or convex lens that are thicker in the mid-
dle like the lenses of our eyes.
Presbyopia, as discussed earlier, is a condition associated with older eyes (or older
people). As we age, the lens hardens and the ciliary muscles lose power. This makes
it harder for older eyes to accommodate to nearby objects. Presbyopia increases the
distance of the near point, which is the closest object we can bring into focus. As with
hyperopia, the lens now focuses objects behind the retina, requiring eyeglasses to be fit-
ted to allow magnification, which will focus objects onto the retinae. But the difference
is that in hyperopia, one corrective lens will allow the person to see near and far but, in
presbyopia, the person needs different corrections for near and far. Thus, your parents
might have reading glasses scattered around the house to use when reading but not when
looking in the distance, or they might wear bifocals where they look through the top for
seeing in the distance and the bottom to see up close. Presbyopia is a seemingly universal
condition. It is seldom noticed before the age of roughly 40. But by the age of 50, there
are few people who do not need reading glasses to correct presbyopia. Indeed, after 50
years of age, the rate at which presbyopia advances might be quite steep, requiring fur-
ther magnification in order for individuals to see small objects close to their eyes.
Astigmatism
Astigmatism develops from an irregular shape of the cornea or an irregular shape of the lens,
which makes it impossible for the lens to accommodate a fully focused image. Usually the lens
will accommodate to a relatively small area of the visual world. In the case of the cornea, the
surface of the cornea may be asymmetrical, and in the case of the lens, the shape of the lens
may not be sufficiently spherical. In astigmatism, the cornea bends the light more strongly in
one direction, say in the vertical direction, than it does in the direction that is at a right angle to
the first direction, say the horizontal direction. Thus, no matter how the lens changes shape, an
object in a particular location and orientation will not be in focus (Figure 3.32). Astigmatism
makes seeing some orientations blurry, whereas others are still relatively clear. Look at ISLE
3.14 for an example of what it looks like to have astigmatism. To correct this problem, a

80 Sensation and Perception
complex, conelike lens is used to offset the differences, though
it is often difficult to correct an astigmatism with eyeglasses
alone. Sometimes contact lenses are used to directly correct the
problem. More recently, LASIK (laser-assisted in situ ker-
atomileusis) surgery is used to reshape the cornea to bend the
light in the correct fashion.
Cataracts
Cataracts result from the clouding of the lens. Cataracts
affect older adults more than they do younger adults, but they
can occur in anyone at any age. With cataracts, over time,
the clear appearance of the lens becomes cloudy because of
water buildup, eventually leading to blindness unless there is
surgical intervention. Cataracts may result from complica-
tions of diabetes, exposure to ultraviolet radiation, or sim-
ply natural aging. Cataracts can now be fixed surgically. The
biological lens is removed and replaced by an artificial lens.
After surgery, patients will need glasses to focus at particular
distances, but they can see normally with glasses.
The conditions we have discussed so far arise from
problems with the lens or cornea. However, there are also
some conditions that affect the retina itself. These problems tend to be far more tragic
than most problems with the lens, which are now treatable with eyeglasses or, in the case
of cataracts, with a relatively straightforward surgical procedure.
Macular Degeneration
Macular degeneration is a disease that destroys the fovea and the area around it. The
medical term macula refers to the center of the retina. The macula includes the fovea but
is larger than it. The destruction of the macular region causes a blind spot in central vision.
A patient will direct his or her point of gaze to a point in space and then promptly lose
sight of it. Macular degeneration can occur in younger adults but is more common in older
adults. Macular degeneration occurs in two forms, known as wet and dry. Wet macular
degeneration has a very fast onset but is partially treatable. Dry macular degeneration may
take years to develop, but no treatment is available at present. Wet macular degeneration
occurs because of abnormal growth of blood vessels, leading to the leaking of blood below
the retina. This leakage leads to scarring of the macula. Wet macular degeneration can
be treated with a variety of medicines that shrink the damaged blood vessels, leading to
some improvement of vision. However, injections must be given frequently to maintain
this improved visual function (Arroyo, 2006). Wet macular degeneration is the less com-
mon form of macular degeneration (Bressler, 2004). Dry macular degeneration, the more
common form, results from degeneration of the cells that produce photopigments (pigment
epithelium) for the cones (and rods) in the macular region of the retina. This results in
impaired function of those photoreceptors (Stone, 2007). No treatment is yet available for
the dry form of macular degeneration. Figure 3.33 in the middle photograph illustrates
what the visual world looks like to patients with macular degeneration. Figure 3.34a is an
ophthalmoscopic fundus photograph of a patient with macular degeneration. You can also
go to ISLE 3.15 for an interview with a person living with this vision disorder.
Retinitis Pigmentosa
Retinitis pigmentosa is an inherited progressive degenerative disease of the retina
that may lead to blindness. Over a long period of time, retinitis pigmentosa leads to
Cornea—sphere shape
Cornea—oval shape
One focal
point
Multiple
focal points
Astigmatism
XX
Normal
vision
X
(a)
(b)
FIGURE 3.32 Astigmatism
(a) A normal cornea. (b) A cornea with an irregular shape,
allowing astigmatism to develop.
©
C
an
S
to
ck
P
ho
to
In
c.
/a
lil
a
Cataracts: a condition that
results from a darkening of the
lens
Macular degeneration: a
disease that destroys the fovea
and the area around it
Retinitis pigmentosa:
an inherited progressive
degenerative disease of the
retina that may lead to blindness

81 Chapter 3: Visual System: The Eye
ISLE 3.14
Astigmatism
ISLE 3.15
Living With Macular
Degeneration and
Retinitis Pigmentosa
FIGURE 3.33 Seeing With Macular Degeneration and Retinitis Pigmentosa
Macular degeneration affects central vision, blurring our perception of whatever it is we are trying to look at. Retinitis pigmentosa
affects peripheral vision, creating a “tunnel” vision effect in which we see what we are looking at with normal acuity but cannot see
what surrounds our focus.
©
C
or
de
lia
M
ol
lo
y/
Sc
ie
nc
e
So
ur
ce
degeneration and destruction of the photoreceptors, particularly rods at the periph-
ery. Indeed, some patients will exhibit symptoms such as night blindness as children,
whereas in other patients, the condition might not develop until adulthood (Hartong,
Berson, & Dryja, 2006). In many cases, the rods are affected first, particularly those
midway through the periphery. This results in tunnel vision (Figure 3.33 on the right),
in which a person can see only in a small area in the middle of the visual field. Retinitis
pigmentosa may progress and cause complete blindness as the destruction of photore-
ceptors spreads to the cones (Figure 3.34b). At present, there is no treatment for this
condition. However, some recent research suggests that gene therapy might be useful
in the near future (Wert, Davis, Sancho-Pelluz, Nishina, & Tsang, 2013). You can see a
person describe his experience with retinitis pigmentosa in ISLE 3.15.
TEST YOUR KNOWLEDGE
1. Examine how differences in the structure of the eye contribute to the differences between
myopia and hyperopia.
2. Choose a refractive error you do not have and predict what vision is like to a person with that
refractive error (it would not hurt to ask someone with this refractive error and see if you are right).
FIGURE 3.34 Macular Degeneration and Retinitis Pigmentosa and the Retina
(a) The retina of a person with advanced dry macular degeneration. (b) The retina of a person with retinitis pigmentosa.
©
B
io
ph
ot
o
A
ss
oc
ia
te
s
/
Sc
ie
nc
e
So
ur
ce
©
S
ci
en
ce
S
ou
rc
e
(a) (b)

82 Sensation and Perception
EXPLORATION: Animal Eyes
Earlier we briefly covered the topic of how our eyes have
resulted from evolutionary processes. From this perspec-
tive, our eyes are the way they are because they helped our
ancestors survive. We have an eye that works for us and
the things that we as humans do. Different animals live in
different ecological niches and, thus, have different eyes.
They use their vision differently to survive in their niches.
There is not space here to cover all the different types of
animal eyes, but we briefly examine a few examples to
show some of the variety in eyes and their visual func-
tions. Some eyes are more complex and some are simpler,
but that is not the same as better or worse. If the eye helps
the animal survive and reproduce, it does its job. So, if the
animal is surviving then that eye works, even if it is not
equally complex to ours. Certainly, we would not enjoy
having any of these other eyes, but then we are humans,
not cats or fiddler crabs.
Cats
Let us start with a familiar animal that, in many ways, has
eyes similar to ours, cats. They are fellow mammals and
have the same basic structure to their eyes and retinae. The
similarity of their eyes to ours has made them useful sub-
jects for learning about our visual systems. Kuffler (1953)
studied cats to discover center-surround receptive fields.
Cats can see well in both daylight and dark. In particular,
they function in the dark much better than humans do.
This need to function in the dark has led to some very
interesting adaptations in their eyes. First, notice the pupil
of cats (Figure 3.35). Instead of a round pupil like ours,
they have a slit pupil, like snakes. This type of pupil allows
them to adjust the size of the pupil to a much greater
degree than we can. Figure 3.35a shows a small pupil for
daytime and Figure 3.35b shows a large pupil for seeing in
the dark. You can barely make out the iris in Figure 3.35b.
This greater ability to change pupil size helps cats function
well in both daylight and nighttime, but it comes at a cost.
This pupil shape does not help make as clear an image as
our circular pupil, so cats have poorer acuity than we do
(Blake, Cool, & Crawford, 1974).
Another way that cats are adapted to see in the dark is
that they do not have a fovea. They have an area cen­
tralis, which is related to their direction of gaze but
does not support high acuity as our fovea does. Another
interesting difference in cat eyes relative to ours has to
do with the layer of cells right behind the receptors.
Our pupils look black because behind our receptors we
have the pigment epithelium (review Figure 3.7). This
layer captures most of the light that is not caught by
a receptor, preventing light scatter, which would make
our retinal image blurrier. Cats, needing to function well
at night, need to get all of the light they can. Instead of
the pigment epithelium, they have a tapetum, which is
a reflective layer bouncing light not caught by receptors
back into the retina (Figure 3.36). It is a tapetum in the
eyes of a deer, another animal that needs to see at night,
that leads to the glow when a “deer is caught in the
headlights.”
FIGURE 3.35 Cat’s Slit-Shaped Pupil
(a) The pupil closed for seeing in bright lights. (b) The pupil opened to assist seeing in the dark.
(a) (b)

83 Chapter 3: Visual System: The Eye
Nautiluses
Cats’ eyes are recognizable to us because of the simi-
larity of their structure and their use, but some eyes are
strange indeed. Let us consider the nautilus. The nautilus
is a marine mollusk (Figure 3.37a) that seems not to have
significantly evolved during the past 500 million years.
To give some context, the very first dinosaurs emerged
230 million years ago and became extinct 66 million
years in the past. Nautliuses live in deep water with lim-
ited light and they are mostly scavengers, so their food
does not try to escape them. There is not a lot of light in
deep water, and the visual detection task of getting food
is not very complicated.
The nautilus has a very unusual eye that could even be
called a primitive eye, a pinhole eye. It does not have a
lens or cornea at all, and its pupil is quite small. Thus,
the nautilus eye is open to the sea (Figure 3.37b). This
is a simple eye, but a pinhole can focus images (Nilsson,
1989). Figure 3.38 shows several images of a partial
eclipse of the sun seen through the holes of a colander.
The small holes, similar to a pinhole pupil, can focus
an image somewhat similarly to a lens. Each hole in
the colander created its own image of a crescent sun.
However, for pupils to focus clearly, they must be small
and, as such, let in only a little light. So, this way of
seeing is only good when you have plenty of light. The
eyes of nautiluses stayed simple while the octopuses
and squids, to whom they are closely related, evolved
eyes that are similar to those of mammals in their
complexity.
Arthropods (insects and crustaceans) have compound
eyes. We will use the fiddler crab as an example but
emphasize the common features of their compound
eyes (Alkaladi & Zeil, 2014). In the compound eye,
FIGURE 3.36 The Reflective Tapetum Layer
The green seen in the cat’s eye in this flash photograph shows
light reflecting back from the tapetum.
FIGURE 3.37 The Nautilus and Its Eye Without a Lens
(a) A photograph of a nautilus. (b) A cross section of the eye.
Notice the lack of a lens and how the eye is open to the water.
©
S
hu
tt
er
st
oc
k.
co
m
/w
ild
es
ta
ni
m
al
(a)
(b)

84 Sensation and Perception
there is no single entrance to the eye but many small
structures called ommatidia (ISLE 3.16). The surface of
the eye is then a complex faceted structure with each
facet being the entrance to a different ommatidium.
Each ommatidium has its own small lens and a small
collection of pigment cells that capture the light and do
transduction. In addition, they have additional pigment
cells that, like our pigment epithelium, limit light scat-
ter, but in this case keep light from passing to adjacent
ommatidia (Figure 3.39). These different ommatidia
have their own neural connections, but they also inter-
act similar to light falling on one cone inhibiting light
responses to adjacent
cones through horizon-
tal cells. As such, these
eyes have lateral inhi-
bition. In fact, one of the very early studies of lateral
inhibition is in the compound eye of the limulus or
horseshoe crab (Hartline & Ratliff, 1958). Compound
eyes are particularly good at detecting motion. Given
the size of the ommatidium relative to our receptors,
motion of objects creates a flicker effect as adjacent
ommatidia turn on and off. As we will see in Chapter
8, motion detection is a very important task for our
visual systems.
ISLE 3.16 Compound Eyes
FIGURE 3.39 Compound Eye of the Fiddler Crab
(a) A photograph of a fiddler crab. Note the eyes at the end of the stalks. (b) A diagram of a representative compound eye showing a
few of the ommatidia.
©
S
hu
tt
er
st
oc
k.
co
m
/N
ic
ho
la
s
To
h
FIGURE 3.38 Light Passing Through the Small
Pinholes of a Colander
Light passing through the small pinholes of a colander. Each hole is a
pinhole and each pinhole focuses a separate image of the solar eclipse.
The picture was taken August 21, 2017, during the Great American
Eclipse. This is the partial eclipse on the way to totality.
TEST YOUR KNOWLEDGE
1. Examine how differences in the structure of the eyes of different animals alter their
experience of the visual world.
2. Formulate how the structure of an animal’s eyes, including ours, might have evolved to
match the animal’s needs to survive.
Tapetum: a reflective layer behind the receptors of
nocturnal animals that bounces light not caught by
receptors back into the retina
Compound eye: an eye that does not have a single entrance
but is made up of many separate components called ommatidia
(a) (b)

85 Chapter 3: Visual System: The Eye
APPLICATION: Vision Prostheses
Vision prostheses are mechanical devices intended to
restore visual function to blind individuals. They are
sometimes referred to as bionic eyes. Unlike the Bionic
Woman of television fame, vision prostheses are real and
are being developed to restore vision to those who have
become blind for a number of reasons. In particular,
they may be of use in people with macular degeneration
and retinitis pigmentosa, diseases that affect transduc-
tion in the eye. Blindness due to brain damage will not
be improved by vision prostheses. Moreover, although
vision prostheses are potentially useful to people who
have lost vision, at present, they are not helpful to con-
genitally blind individuals.
In the Western world, macular degeneration and retinitis
pigmentosa are the most common forms of blindness. In
both conditions, the photoreceptors are damaged, but
the underlying retinal ganglion cells and the optic nerve
remain intact. Therefore, the goal of vision prostheses
is to replace the missing photoreceptors with essentially
an artificial layer of photoreceptors, sometimes called an
“artificial retina.” There are a number of developments
in artificial retinae, but only one company has market-
ing approval for its device from the U.S. Food and Drug
Administration, granted in 2013, having previously
received approval in Europe in 2011. It is called the
Argus II, developed by California-based Second Sight
Medical Products (McKeegan, 2013). At present, it is
approved only for patients with profound retinitis pig-
mentosa. See Figure 3.40 for an illustration.
Unlike the Bionic Woman, the Argus II is still a bit bulky.
A small video camera is mounted on a pair of eyeglasses
the patient wears. Images from the video camera are
relayed wirelessly to a chip placed directly (via surgery)
on the patient’s retina or retinae. The chip’s electrodes
then stimulate the retinal ganglion neurons in a pat-
tern that roughly approximates the image the camera
is detecting. In its current form, the Argus II creates a
visual field approximately 20 degrees wide, thus allow-
ing some central vision, though with only limited periph-
eral vision. In the current version of the Argus II, there
are only 200 electrodes in the chip. It is estimated that
it would take 1,000 of these to allow the patient to see
clearly enough to recognize individual faces (Humayun
et al., 2012). Thus, at present, this system also has
limited visual acuity. It is likely that at present, reading is
also still not an option. But as the technology improves,
more electrodes may be loaded onto the chip, allowing
enhanced acuity.
Interestingly, patients don’t immediately “see” again.
They have to be trained to interpret the signals they
are receiving. However, with just a few hours’ worth
of training, patients can avoid obstacles, locate objects,
recognize colors, and even recognize very large letters.
Obviously, it is far from perfect. Although patients can
localize objects, tests of visual acuity show that they
improve to, at best, only approximately 20/1,000.
Moreover, the camera is not stable relative to the eyes,
and patients have to keep up with these small move-
ments of the camera relative to their eyes. It appears
they use their auditory system to assist in these needed
adjustments (Barry & Dagnelie, 2016). Nonetheless,
many patients are pleased to have some vision restored.
You can see a video of a person with the Argus II in
ISLE 3.17.
The Argus II is not the only device being built that uses
remote cameras to directly stimulate the optic nerve.
A number of other companies are developing similar
devices. One device, the artificial silicon retina, implants
the visual sensors directly into the damaged retina
(Chow et al., 2004). It then electrically stimulates the
retinal ganglion cells (DeMarco et al., 2007). This device
has the advantage of not requiring cameras and external
electronics. However, at present, clinical trials are still
ongoing, and the device is not yet approved for general
use. Another interesting approach to restoring eyesight
in those with retinal disease involves regrowing photo-
receptor cells using stem cell therapy. This research is
still in its beginning stages but is showing some prog-
ress. Other researchers are using fetal retinal tissue to
regrow photoreceptors in patients with retinitis pigmen-
tosa. One study found improvements in patients’ vision
up to 5 years later. One patient improved from 20/800
to 20/200 acuity after tissue implants (Radtke et al.,
2008). Like other techniques, this one is still in the clin-
ical trials stage.
There is also one device, BrainPort, that tries to use tactile
senses on the tongue to help those who are blind. As with

86 Sensation and Perception
the Argus II, patients wear glasses with a camera. However,
there is no implant. The camera goes to a controller that
converts the visual signals from the camera to tactile signals
that travel to a pad worn on the tongue. The pad vibrates
a pattern to match the visual information, and people can
learn to interpret this information to allow them to “see” in
much the same way as the blind can read using their fingers
in Braille (Grant et al., 2016).
Although wearing a pad in
the mouth all the time has its
limitations, this type of visual
prosthesis will work for those who have damage in the
visual cortex. Go to ISLE 3.17 for more about BrainPort,
including a video of people using this device.
TEST YOUR KNOWLEDGE
1. Apply your knowledge of center-surround receptive
fields to the challenges of building successful visual
prostheses.
2. Evaluate if visual prostheses would have to perform light
and dark adaptation.
Electronics Case
Antenna
Processor
Electrode Array
Camera
Antenna
FIGURE 3.40 Vision Prostheses
In the Argus II, images are created by a video camera mounted
on eyeglasses worn by the patient. The images are then sent to
a computer processor on the eyeglasses themselves. The signal
is then sent wirelessly to a chip embedded on the retina, which
then stimulates the retinal ganglion cells.
CHAPTER SUMMARY
3.1
Appraise the nature of light as electromagnetic
energy to explain the relation of wavelength to
perception.
Visible light is a small slice of the electromagnetic spectrum.
The electromagnetic spectrum is the complete range of
wavelengths of light and other electromagnetic energy. Light
from a light source such as the sun or a reading lamp reflects
off objects in the environment and enters our eyes. The pat-
tern of light enters our eye, and then our visual processes
start decoding it. Light is made up of particles called photons
that also behave in a wavelike manner. We see differences
in wavelengths as different colors.
3.2
Sketch the anatomy of the eye to show how it
makes vision possible.
The cornea and the pupil control the amount of light that
enters the eye. The size of the pupil adjusts to control the
amount of light entering the eye, and then the lens focuses
that light on the retina to make a retinal image. The retina
contains photoreceptors, which convert the light energy into
a neural signal.
3.3
Interpret how the retinae transduce light energy
into a neural signal that is sent to the brain.
The rods and cones of the retina are able to convert light
into a neural signal through chemicals known as photopig-
ments. Photopigments are molecules that absorb light and by
doing so release an electric potential by altering the voltage
in the cell. When a photopigment absorbs a photon of light,
it changes shape. This change of shape initiates a series of
biochemical processes, which result in a change of electric
potential, which allows for a signal to exit the photoreceptor.
In this way, a neural signal leaves the photoreceptor and is
transmitted to the optic nerve to be sent to the brain. Opsin is
the protein portion of a photopigment that captures the photon
of light and begins the process of transduction. It is the vari-
ation in this opsin that determines the type of visual receptor.
3.4
Illustrate how the visual system adapts to the
dark and then adapts again to the light to help us
adjust to changing light levels.
The retina has two types of photoreceptors, rods and cones.
The rods are specialized for night vision (scotopic) and light
ISLE 3.17
Vision Prostheses

Chapter 3: Visual System: The Eye 87
detection, whereas the cones are specialized for visual acuity
(photopic) and color vision. The duplex theory of vision states
that there are functionally two distinct ways that our eyes
work, one, the photopic, associated with the cones, and the
other, the scotopic, associated with our rods. Dark adaptation
takes about 15 minutes. Technologies have been developed to
allow people to see visual displays, such as computers with
red light that do not interfere with dark adaptation. The rods
and cones project to the retinal ganglion cells.
3.5
Diagram the receptive fields of retinal ganglion
cells, and show how they need contrast to
operate.
The rods and cones project to cells called the retinal ganglion
cells, whose axons exit the eye through the optic nerve. Retinal
ganglion cells have receptive fields. Receptive fields are regions
of adjacent receptors that will alter the firing rate of cells that
are higher up in the sensory system. The term can also apply to
the region of space in the world to which a particular neuron
responds. These receptive fields have a characteristic shape
known as center-surround. Center-surround receptive fields
occur when the center of the receptive field responds opposite
to how the surround of the receptive field responds. If the center
responds with an increase of activity to light in its area, the sur-
round will respond with a decrease in activity to light in its area
(on-center). If the center responds with a decrease of activity
to light in its area, the surround will respond with an increase
in activity to light in its area (off-center). This creates the phe-
nomenon of lateral inhibition. Lateral inhibition is the reduction
of a response of the eye to light stimulating one receptor by
stimulation of nearby receptors. It is caused by inhibitory sig-
nals in horizontal cells and creates the Mach band illusion. It is
responsible for enhanced edge detection by our visual system.
3.6
Judge the different refractive errors and eye dis-
eases on how they affect the ability of the eye to
see.
We also described a number of diseases and problems with
the eye. Myopia, presbyopia, and hyperopia are common
problems in the refracting power of the lens. Astigmatism is
a problem in the shape of the lens or cornea. They are all cor-
rectible with eyeglasses. Cataracts occur when the lens gets
cloudy and allows less or no light to pass through. Macular
degeneration and retinitis pigmentosa are diseases of the ret-
ina that can lead to serious visual impairment. Macular degen-
eration leaves a blind spot right at the location we are trying
to look at. Retinitis pigmentosa can cause tunnel vision in its
early and middle phases and lead to total blindness later on.
REVIEW QUESTIONS
1. What is light? Why is it considered both a wave and
a particle?
2. What are the roles of the cornea, the iris, the pupil,
and the lens in human vision?
3. What is accommodation? Explain how it produces a
focused image on the retina.
4. What are the different classes and subclasses of
photoreceptors? How do they convert light into a
neural signal?
5. What is the difference between photopic vision and
scotopic vision? What is each used for, and what
physiological mechanism underlies each?
6. What is the Purkinje shift? What does it tell us about
photopic vision and scotopic vision? How does it
relate to dark adaptation?
7. What are retinal ganglion cells? What do they do in
visual processing?
8. What are center-surround receptive fields? What
is the difference between on-center and off-cen-
ter center-surround receptive fields? How do these
explain lateral inhibition?
9. Distinguish between myopia, presbyopia, and hyper-
opia. What are the differences between macular
degeneration and retinitis pigmentosa?
10. What are some of the ways that eyes vary across
different animals?
11. What are vision prostheses? How do they restore
vision to people with retinal disease?

Sensation and Perception88
PONDER FURTHER
1. It is often written that the eye is like a camera, but that
analogy is not quite correct. You have read a lot about
the eye in this chapter. Create a diagram that compares
the eye to a camera, and consider what functions are
similar and where the analogy breaks down. Consider
both light-gathering elements of the eye and what is
done after light is captured, that is, after transduction
in the eye and after light falls on the film in the camera.
2. Review ISLE 13.11 (both parts) and ISLE 13.12. The ear-
lier discussion suggests that these receptive fields play
a role of edge enhancement. Survey graphic designs
around you, on the web, in print ads, or anywhere else,
and consider how these ads seem to take advantage of
this edge enhancement feature of our eye.
KEY TERMS
Accommodation, 60
Anterior chamber, 60
Astigmatism, 79
Cataracts, 80
Center-surround receptive field, 74
Ciliary muscles, 61
Compound eye, 83
Cones, 64
Convergence, 70
Cornea, 59
Dark adaptation, 70
Duplex theory of vision, 68
Edge detection, 74
Electromagnetic energy, 55
Electromagnetic spectrum, 56
Field of view, 58
Fovea, 64
Frequency, 55
Heterochromia, 60
Hyperopia, 78
Hyperpolarization, 67
Intensity, 55
Iris, 60
Lateral inhibition, 76
Lens, 60
Light adaptation, 71
Macula, 64
Macular degeneration, 80
Myopia, 78
Near point, 61
Neurotransmitter, 67
Off-center receptive fields, 75
On-center receptive fields, 75
Opsin, 66
Optic disc, 65
Photon, 56
Photopic vision, 68
Photopigment, 66
Posterior chamber, 60
Presbyopia, 61
Pupil, 60
Pupillary reflex, 60
Purkinje shift, 69
Receptive field, 73
Retina, 62
Retinal, 66
Retinal image, 62
Retinitis pigmentosa, 80
Rods, 64
Sclera, 60
Scotopic vision, 68
Tapetum, 82
Wavelength, 55
Zonule fibers, 61

Chapter 3: Visual System: The Eye 89
Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e
SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review,
study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning.
Learning Objectives Digital Resources
3.1 Appraise the nature of light as electromagnetic energy to
explain the relation of wavelength to perception.
Light
3.2 Sketch the anatomy of the eye to show how it makes vision
possible.
The Eye Pupil’s Response to Static and Dynamic Illusions of
Luminosity and Darkness
3.3 Interpret how the retinae transduce light energy into a neural
signal that is sent to the brain.
2-Minute Neuroscience: The Retina
Bionic Eye Opens New World of Sight for Blind
3.4 Illustrate how the visual system adapts to the dark and then
adapts again to the light to help us adjust to changing light levels.
The Eye Pupil Adjusts to Imaginary Light
3.5 Diagram the receptive fields of retinal ganglion cells, and show
how they need contrast to operate.
On and Off Center Retinal Cells
3.6 Judge the different refractive errors and eye diseases on how
they affect the ability of the eye to see.
The Manital Cortex in the Blind: Lessons About Plasticity
and Vision
Darkness Provides a Fix for Kittens With Bad Vision

4Visual System: The Brain
Richard Kail/Science Source
LEARNING OBJECTIVES
4.1
Identify the anatomy of the optic chiasm and how that
affects the lateralization of vision in the brain.
4.2
Diagram the anatomy of the lateral geniculate nucleus and the
superior colliculus, and describe their roles in visual processing.
4.3 Explain the nature of the retinotopic organization of V1 and the organization of V2.
4.4 Compare the difference between the dorsal pathway and the ventral pathway.
4.5 Interpret the concepts of blindsight and conjugate gaze palsy.
INTRODUCTION
In 1981, two Harvard professors won the Nobel Prize in Physiology or Medicine for
their work on the anatomy and physiology of the visual cortex (Figure 4.1). The two
men, Torsten Wiesel (born 1924) and David Hubel (1926–2013), collaborated for more
than 20 years on research that has revolutionized our understanding of the mammalian
visual cortex, including the human visual cortex. Dr. Wiesel, originally from Sweden,
and Dr. Hubel, born in Canada to American parents, started their collaboration as
postdoctoral fellows with Stephen Kuffler (whose work was discussed in Chapter 3).
A postdoctoral fellow, or postdoc, is someone who has earned his or her PhD but is
learning new techniques in a senior researcher’s laboratory. Later, they both moved to
Harvard and continued working together as they rose through the ranks at Harvard.
Their contribution to the field of sensation and perception cannot be overstated. We
refer to work they did 50 years ago, and it is just as relevant today as it was when it was
first done. Both men continued to be active professionally well into their 80s. Indeed,
in 2005, they published a book together outlining their collaborative research over
the decades, including much work done after their Nobel Prize was awarded (Hubel
& Wiesel, 2005). Wiesel also has been active in promoting global human rights and in
promoting collaboration between Israeli and Palestinian scientists.
Among their amazing discoveries is that areas of the visual brain, including
areas in the thalamus and areas in the occipital cortex, are specifically sensitive to
certain kinds of stimuli and that this sensitivity can be mapped into predictable
patterns (Hubel & Wiesel, 1965). That is, they discovered neurons in the brain
that have specific visual fields; the cells respond to some patterns of stimuli but not
ISLE EXERCISES
4.1 From Eye to LGN
4.2 Simple Cells
4.3 Complex Cells
4.4 Hypercolumns
4.5 Navigation in
Blindsight

92 Sensation and Perception
others. They also found that these stimulus-specific cells are
organized into complex but predictable columns. In related
research, they also pioneered how environmental input
affects the development of the mammalian nervous system,
that is, how these systems grow and change as mammalian
babies get older.
In this chapter, we consider the visual brain, namely, the
networks in the lateral geniculate nucleus (LGN) of the thal-
amus and the networks in the occipital cortex that allow
us to see in the way we do. It is these areas of the brain
that allow us to perceive beautiful sunsets, drive very fast
cars at racetracks, and quietly read at night by the glow of
our tablet computers. However, at the outset, we issue our
readers a warning: Do not read this chapter once and think
you understand the material. The organization of the visual
system is complex, and our derived naming system for it can
tax our mnemonic abilities—you will encounter parasol ret-
inal ganglion cells, parvocellular layers, hypercolumns, and
one area of the brain referred to by five different names.
Therefore, our advice is to review repeatedly, do all of the online demonstrations,
and self-test as you go through the material in this chapter. You will be glad you
did—the organization of the human visual system is a wonder of the natural world,
one to which science has paid particularly close attention.
In Chapter 3, we discussed how light is transduced by the retinae of the eyes
and converted into a neural signal. The retinae begin the processing of the visual
signal, already looking for edges and coding for color. This process intensifies when
the optic nerves leave the eyes and head for their first synapses in the brain. In this
chapter, we follow the neural signal as it leaves the eye and travels first to the LGN
of the thalamus and from there to the visual cortex in the occipital lobe. We also
consider other pathways of visual information after it leaves the eye. Much of the
information we consider in this chapter comes from research on nonhuman eyes
and brains, as most of the electrophysiological research was done on rats, cats, and
monkeys. However, much of what we know about mammalian visual systems gener-
alizes to our own. At the end of this chapter, we also consider a neuropsychological
condition called blindsight, which results from damage to the visual regions of the
brain without damage to the eyes. We will also consider conjugate gaze palsy, which
arises from damage in the brain stem. Both of these peculiar conditions underline the
critical nature of the brain in our ability to see.
THE OPTIC NERVE AND CHIASM
4.1
Identify the anatomy of the optic chiasm and how that
affects the lateralization of vision in the brain.
Approximately 1 million retinal ganglion cells form the optic nerve of each eye. The optic
nerve of the left eye and the optic nerve of the right eye meet just a couple of centimeters
behind the eyes, in an area called the optic chiasm. Here at the optic chiasm, your visual
system does one of its most interesting tricks, which is also one of the most difficult
©
Ir
a
W
ym
an
/S
yg
m
a/
Co
rb
is
FIGURE 4.1 David Hubel and Torsten Wiesel
David Hubel and Torsten Wiesel in 1982, shortly after
winning the Nobel Prize in 1981 for their work on the visual
system.
ISLE 4.1
From Eye to LGN
Optic chiasm: the location in
the optic tract where the optic
nerve from each eye splits in
half, with nasal retinae crossing
over and temporal retinae
staying on the same side of the
optic tract
Optic tract: the optic nerve
starting at the optic chiasm and
continuing into the brain
Contralateral representation
of visual space: the
arrangement whereby the left
visual world goes to the right
side of the brain, and the right
visual world goes to the left side
of the brain
Ipsilateral organization:
same-side organization; in the
visual system, the temporal
retina projects to the same side
of the brain

93 Chapter 4: Visual System: The Brain
aspects of visual anatomy for people new to the area to
keep straight (Figure 4.2). So again, we warn you to pay
close attention and review this information multiple times.
The optic nerve from each eye splits in half at the
optic chiasm. The axons from the ganglion cells from
the right half of the right retina and the ganglion cells
from the right half of the left retina combine, forming
the right optic tract, which then proceeds to the right
hemisphere of the brain. Axons from the ganglion cells
in the left half of the right retina and the left half of the
left retina combine, forming the left optic tract, which
then proceeds to the left hemisphere of the brain. For an
illustration of how the optic nerve divides in the optic
chiasm, see ISLE 4.1.
Why is this organization important for the visual
system? Consider the right optic tract. It combines ret-
inal signals from the temporal side (near the temple
or toward the forehead) of your right retina with ret-
inal signals of the nasal side (toward the nose) of your
left retina. If you examine Figure 4.2, you will see that
these two retinal areas receive input from the left visual
world. Similarly, the left optic tract receives information
from the right visual world. Thus, initially, the brain is
respecting the outside visual world in terms of its rep-
resentation in the brain. The left world goes to the right
hemisphere, and the right world goes to the left hemi-
sphere. This organization is known as contralateral rep-
resentation of visual space.
We wish to reiterate this point, as it serves to dispel
some myths about the eyes and the brain. Here’s the
main point of this paragraph: (a) Information from each
eye goes to both hemispheres, and (b) each hemisphere
of the brain receives input from the contralateral visual
field. So the left eye sends information to both the left
and right hemisphere. However, the left visual field of
the world is initially sent to the right hemisphere. For
example, if you consider the left eye only, the left tem-
poral retina receives information from the right visual
world, which stays on the same side of the brain and
goes to the left hemisphere (ipsilateral organization, or
same-side organization). The left nasal retina receives
information from the left visual world, which crosses
over and goes to the right hemisphere (contralateral
organization). Thus, each eye sends a signal to each
hemisphere. And it is the visual world that projects to
the contralateral side of the brain. We know: It sounds
confusing. Study Figure 4.2 closely, as in this case, the picture better conveys this
organization.
Once the optic tract has left the chiasm, 90% of the axons make their way to
the LGN of the thalamus. Here, the organization continues to respect the left-field/
Contralateral organization:
opposite-side organization; in
the visual system, the nasal
retina projects to the opposite
side of the brain
FIGURE 4.2
The Optic Chiasm and Visual Space in the Brain
(a) The optic nerves leave the eyes and travel toward the brain. At the
optic chiasm, each optic nerve splits apart. Information from the left
visual world first travels toward the right side of the brain, whereas
information from the right visual world first travels toward the left side of
the brain. Information from each eye, however, goes to both sides of the
brain. For example, the left temporal retina receives information from the
right visual world, which stays on the same side of the brain and goes to
the left hemisphere (ipsilateral or same-side organization). The left nasal
retina receives information from the left visual world, which crosses
over and goes to the right hemisphere (contralateral organization).
(b) Use your anaglyph glasses to see the optic chiasm as it appears in
the sheep brain. It is nearly identical to how it appears in the human brain.
Optic nerve
Optic chiasm
Lateral
geniculate
nucleus
Visual cortex
A
A
A A
B
B
B B
(a)
(b)

94 Sensation and Perception
right-field distinction. This pathway then leads to the visual cortex. However, 10% of
the axons from the optic tract go to other locations in the brain. Many of these axons
that do not go to the thalamus go instead to a locus in the brain known as the superior
colliculus, a midbrain structure that sits below the LGN. The superior colliculus is an
important structure in eye movements. A small number of these axons also head to
the frontal eye field region—an area in the frontal lobe also instrumental in eye move-
ments. Some fibers also go to the pineal gland, which regulates our circadian rhythms.
It is for this reason that exposure to bright lights can help “reset” our internal clocks.
We will first consider the main pathway from the retinae to the brain.
TEST YOUR KNOWLEDGE
1. What is meant by the term contralateral representation of visual space? Why would the
visual system represent visual space? What function does it serve?
2. Describe how information moves through the optic chiasm. How does this lead to left and
right field organization in the brain?
THE LATERAL GENICULATE NUCLEUS
4.2
Diagram the anatomy of the lateral geniculate nucleus and the
superior colliculus, and describe their roles in visual processing.
The lateral geniculate nucleus is a bilateral structure (one is present in each hemi-
sphere) in the thalamus that relays information from the optic nerve to the visual cor-
tex. The thalamus is a large structure that serves as a relay station for a number of
sensory systems, including both vision and audition (we discuss the medial geniculate
nucleus, critical in hearing, in Chapter 11). The LGN is the critical locus in the thalamus
for vision. We now know that the LGN has complex functions and is not simply a relay
station but is already doing visual processing. However, it is often thought of as a relay
center because it holds an intermediate position between the retinae of the eyes and the
visual cortices.
Anatomically, the LGN is divided into six layers (Figure 4.3). Layers 1 and 2 are
called the magnocellular layers because the cells in these layers are large. Layers 3
through 6 are called the parvocellular layers because the cells are somewhat smaller.
The very thin layers between each of the two magnocellular levels and the four par-
vocellular layers are called koniocellular layers, and they consist of very small cells.
Thus, there are six koniocellular layers, one under each of the two magnocellular
layers and each of the four parvocellular layers.
LGN anatomy gets complicated very quickly. Please read carefully, review what
you just read, and then write it down, so you can be confident that you have under-
stood and retained this material. You should also study the figures to reinforce your
understanding of the organization of the LGN.
First, keep in mind that we have two LGNs, one in each hemisphere. The left
LGN receives input primarily from the right visual world, and the right LGN receives
input primarily from the left visual world. Also remember that each eye projects
inputs to both the left and right LGNs when half of the retinal axons cross over in
the optic chiasm. Once in the brain, it is the representation of the visual world that is
more important than which eye this information came from (except when consider-
ing depth perception). This is important to repeat, because it is often misrepresented
outside of scientific publications.
Lateral geniculate nucleus:
a bilateral structure (one is
present in each hemisphere)
in the thalamus that relays
information from the optic
nerve to the visual cortex
Magnocellular layers:
layers of the lateral geniculate
nucleus with large cells that
receive input from M ganglion
cells (parasol retinal ganglion
cells)
Parvocellular layers: layers
of the lateral geniculate
nucleus with small cells that
receive input from P ganglion
cells (midget retinal ganglion
cells)
Koniocellular layers: layers
of the lateral geniculate
nucleus with very small cells
that receive input from K
ganglion cells (bistratified
retinal ganglion cells)

95 Chapter 4: Visual System: The Brain
FIGURE 4.3 The Lateral Geniculate Nucleus (LGN)
The LGN is a six-layered structure. Layers 1 and 2 are called the magnocellular layers because the cells in these layers are large.
Layers 3 through 6 are called parvocellular layers because the cells are somewhat smaller.
Cortex
1
2
3
4
5
6
Left LGN
Thalamus
Right LGN
Optic tract
Optic nerve Layers 1 and 2 are the magnocellular layers.
Layers 3–6 are the parvocellular layers.
So here are some important facts about the LGN. First, each LGN layer receives
input from only one eye. Magnocellular layer 1 and parvocellular layers 4 and 6
receive input from the contralateral eye (i.e., the eye on the opposite side of the
head). This is also true for the koniocellular layers underneath each of these layers.
Magnocellular layer 2 and parvocellular layers 3 and 5 receive input from the ipsi-
lateral eye. This is also true for the koniocellular layers underneath each of these lay-
ers (Hendry & Reid, 2000). This means that the LGN preserves information about
where information in the visual world is coming from and which eye is detecting
that information. Obviously, it is critical to know where objects are in the world, but
knowing which eye an image comes from is important in constructing a three-dimen-
sional image.
It is also critical that particular retinal ganglion cells are projecting to par-
ticular layers in the LGN. The magnocellular layers of the LGN receive input
from the parasol retinal ganglion cells (M cells). The parvocellular layers of
the LGN receive input from the midget retinal ganglion cells (P cells). And
the koniocellular layers of the LGN receive input from the bistratified retinal
ganglion cells (K cells).
The last general feature of the LGN that is important to understand is that the
neurons in each layer of the LGN show retinotopic organization. This means that
retinal ganglion cells from adjacent regions of the retina connect to cells in adja-
cent areas of the LGN. This also means that the LGN has an organization oriented
Parasol retinal ganglion cells
(M cells): retinal ganglion cells
that project to the magnocellular
layer of the lateral geniculate
nucleus; they represent 10% of
ganglion cells and possess high
sensitivity to light
Midget retinal ganglion cells
(P cells): retinal ganglion cells that
project to the parvocellular layer of
the lateral geniculate nucleus; they
represent 80% of ganglion cells,
possess low sensitivity to light, and
are sensitive to wavelength
Bistratified retinal ganglion
cells (K cells): retinal
ganglion cells that project to
the koniocellular layer of the
lateral geniculate nucleus; they
represent 10% of ganglion cells,
possess low sensitivity to light,
and are sensitive to wavelength

96 Sensation and Perception
to the visual world, as adjacent areas in
the visual world are picked up by adja-
cent areas of the retina, which are in turn
projected to adjacent areas of the LGN.
To see all this information graphically,
examine Figure 4.4.
To understand the functional signifi-
cance of this complex anatomy, it is nec-
essary to backtrack a bit and examine
different kinds of retinal ganglion cells.
Remember that the retinal ganglion cells
emerge from specific locations along the
retina. Those close to the fovea receive
input from few or just one receptor cell,
and those at the periphery may be collect-
ing input from many more receptor cells.
However, not all retinal ganglion cells are
the same. Indeed, we see the beginning of
three unique pathways that start in the
retinae and continue to the visual cortex.
We consider each of these pathways now.
The parvocellular pathway (or simply
P pathway) is characterized by the retinal
ganglion cells known as midget retinal
ganglion cells (so named because of their
small size). Midget retinal ganglion cells
usually receive input from a single cone
in the fovea of the retina, thus carrying
detailed information necessary for visual
acuity. When stimulated, these cells show
a sustained response; that is, they continue
to fire throughout the time period when a
stimulus is present. These retinal ganglion
cells are sensitive to wavelength (the basis
for color perception). The koniocellular
pathway (or simply K pathway) starts
with bistratified retinal ganglion cells and
projects to the koniocellular layers of the
LGN. These retinal ganglion cells also
receive input from cones. However, there
is more convergence among these cells,
so that they show lower acuity than the
P-pathway cells. But they do show some
role in color vision.
The magnocellular pathway (or sim-
ply M pathway) starts with the parasol
retinal ganglion cells and projects to
the magnocellular layers of the LGN.
These retinal ganglion cells receive
input from many photoreceptors, including both rods and cones. As such, they
have large receptive fields and are sensitive to light but not to color. They have lower
acuity relative to the K-pathway or P-pathway cells. For this reason, the M system is
usually associated with visual functions such as light detection and motion detection.
FIGURE 4.4 Organization Within the LGN
Objects in the left visual world initially go to the right hemisphere, whereas objects
in the right visual world initially go the left hemisphere. This is represented here
in both the eye and the optic chiasm with color coding. Different types of ganglion
cells project to different layers of the LGN. The parvocellular layers receive input
from the midget retinal ganglion cells, whereas the magnocellular levels receive
input from the parasol retinal ganglion cells. The koniocellular layers receive input
from the bistratified retinal ganglion cells.
Note: RGCs = retinal ganglion cells
Fixation point
Left eye Right eye
Objects in left half
of visual field
Objects in right half
of visual field
Temporal retina
Temporal
retina
Optic chiasm
Left LGN
1
2
3
4
5
6
1
2
3
4
5
6
Right LGN
Parvocellular
layer
Magnocellular
layer
Koniocellular
layer
Signals from right
half of visual field
Signals from left
half of visual field
Signals from
midget RGCs
Signals from
parasol RGCs
Signals from
bistratified RGCs
Nasal retina
Parvocellular pathway
(P pathway): a pathway
characterized by the retinal
ganglion cells known as midget
retinal ganglion cells

97 Chapter 4: Visual System: The Brain
Processing in the LGN
The LGN maintains a retinotopic map of the left or
right visual world in each of its layers (Figure 4.5).
Interestingly, each of these layers emphasizes differ-
ent aspects of visual processing. We know this from
single-cell studies that found adjacent cells in the LGN
that respond to visual stimuli that excite adjacent cells
in the retina. However, the receptive fields of LGN
cells are more like retinal ganglion cells than receptors.
LGN neurons have receptive fields that are similar
in characteristics to the retinal ganglion cells, with cen-
ter-surround organization (Xu, Bonds, & Casagrande,
2002). That is, some LGN cells respond maximally
to stimuli that are present in the center of the cell’s
receptive field but absent outside the center (or
reversed). Thus, like the retinal ganglion cells, LGN
neurons show specific responding to edges, spots,
and gratings. As we discussed earlier, edge detection
is critical to perception, because it allows the visual
system to determine where one object ends and the
next object begins.
Originally, the LGN was thought of as a relay
point—a synapse between the retina and the visual
cortex, where the complex processing occurred. But,
as with much of the human brain, the LGN turns
out to be much more complex than that. There are
many feedback loops from the cortex back to the
LGN and many connections from the LGN to other
areas of the brain (Babadi, Casti, Xiao, Kaplan, &
Paninski, 2010). The LGN receives input from the
cortex, from the brain stem, from other loci in the
thalamus, and from within the LGN itself. Indeed,
there are more connections from the cortex back
to the LGN than there are from the LGN to the
cortex. Many of these pathways are only now being
investigated. Thus, much remains to be discovered
about the complex function of the LGN.
As indicated earlier, studies suggest that the dif-
ferent layers of the LGN have different functions.
The magnocellular layers are sensitive to motion, light detection, and sudden changes
in the visual image. The parvocellular level specializes in foveal functions—color,
acuity, texture, and depth. The koniocellular layers also seem to specialize in color.
One obvious question is about how we know this. Most of what we know about
these functions comes from single-cell recording experiments using nonhuman mam-
malian models. We turn to those now.
Many of the single-cell recording experiments on the LGN use rhesus macaques
as the research participants. Their visual brains are similar to our own. Because the
visual system does not require the animal to be conscious while it is looking at stim-
uli, the monkey does not experience pain or discomfort, as it is fully anesthetized
during the procedure. Using single-cell recording on particular neurons, different
stimuli can be presented to the monkey. For example, a cell in the magnocellular
layer may be sensitive to movement in a particular area of the visual world. A cell in
FIGURE 4.5 Organization in the LGN
Stimuli in the world are mapped in a consistent fashion onto each layer
of the LGN. Input from the left visual field goes to the right LGN, and
input from the right visual field goes to the left LGN. Layers 1 and 2 are
the magnocellular layers receiving input, whereas layers 3 to 6 are
parvocellular. The koniocellular layers, not pictured, are in between.
Left LGN
F E D
6
F E D
5
F E D
4
F E D
3
F E D
2
F E D
1
Right LGN
C B A
6
C B A
5
C B A
4
C B A
3
C B A
2
C B A
1
Left eye Right eye
F
F
E
ED DC
CB
B
A
A
A B C D E F
Left visual field Right visual field
Koniocellular pathway (K
pathway): a pathway that
starts with bistratified retinal
ganglion cells and projects to
the koniocellular layers of the
lateral geniculate nucleus
Magnocellular pathway
(M pathway): a pathway that
starts with the parasol retinal
ganglion cells and projects to
the magnocellular layers of the
lateral geniculate nucleus

98 Sensation and Perception
the parvocellular layer may be maximally sensitive to light of a particular wavelength
in a particular region in central vision. In this way, we can map which cells respond
to which kinds of stimuli (Nassi & Callaway, 2009).
We end our discussion of the LGN with one interesting fact. As mentioned earlier,
the LGN is part of the thalamus. During sleep, the thalamus is inhibited by complex
neural circuitry in the brain. Thus, unless very bright lights are shined directly into the
eye, a person’s eyes can be open during sleep, but they will not see, because information
does not leave the LGN. Information is registered on the retina and is sent to the LGN,
where it stops. Inhibition prevents the information from being sent on to visual areas
of the cortex, thus preventing conscious vision during sleep.
TEST YOUR KNOWLEDGE
1. Describe the neuroanatomy of the lateral geniculate nucleus. How do the different
ganglion cells project to the different layers of the LGN?
2. What are the differences between parasol retinal ganglion cells (M cells), midget retinal
ganglion cells (P cells), and bistratified retinal ganglion cells (K cells)? What kind of
information does each one carry?
THE SUPERIOR COLLICULUS
We have two superior colliculi, one on each side of the brain. The superior colliculus
is at the top of the brain stem, anatomically just beneath the thalamus (Figure 4.6).
Like the LGN, the left superior colliculus receives input from the right visual world,
and the right superior colliculus receives input from the left visual world. Its main
function in mammals (including humans) is the control
of rapid eye movements. The superior colliculus is part
of a largely nonconscious system that helps us direct our
eyes toward new or approaching objects.
Approximately 90% of retinal ganglion cells synapse
in the LGN, but about 10% go to the superior colliculus.
It is likely that these retinal ganglion cells are bistratified
retinal ganglion cells (the K pathway) (Foldvari & Chen,
2016). Like the LGN, the superior colliculus receives
feedback from the visual cortex, and the superior col-
liculus also projects to the koniocellular levels of the
LGN (May, 2006). The superior colliculus also projects
to areas of the visual cortex beyond the primary visual
cortex. As with the LGN, one can find a retinotopic map
of the visual world in cells inside the superior colliculus.
Nonetheless, the main pathway is directly from the ret-
inal ganglion cells to the superior colliculus, allowing
the superior colliculus to produce quick eye movements.
The superior colliculus is often implicated in the phe-
nomenon of blindsight, a phenomenon we examine in
depth at the end of the chapter.
To understand the role of the superior colliculus, we
must first discuss the nature of human eye movements.
In general, human eyes can make two kinds of voluntary
eye movements, smooth-pursuit eye movements and
FIGURE 4.6 The Superior Colliculus
This illustration shows the location of the superior colliculi relative
to other brain structures relevant for vision. The superior colliculus
is involved in controlling eye movements.
Eye
Optic nerve
Optic chiasm
Optic tract
Superior
colliculus
Lateral
geniculate
nucleus
Cerebral cortex

99 Chapter 4: Visual System: The Brain
saccades. We will start with smooth-pursuit movements. Think about watching a bird
fly across the sky: Your eyes slowly and smoothly move as the bird moves across your
field of vision. These smooth-pursuit eye movements can be made only when we are
watching a moving object. If you don’t believe it, try it. Try to slowly move your eyes
across a stationary landscape. Have someone watch you. You will find that you cannot
slowly move your eyes. Rather, your eyes jump from one point to another.
The superior colliculus is vital in making smooth-pursuit eye movements. Consider
that when we are looking directly at an object without moving our eyes (fixation),
activity can be registered in the region of the superior colliculus that responds to the
fovea. When the object begins to move, cells adjacent to the foveal area in the superior
colliculus become active. This helps guide our smooth-pursuit movement, which we
can do only in response to a moving object. There are also close connections between
the superior colliculus and the innervation of the muscles that control our eyes.
The rapid movement of the eye from one stationary object to another is called a
saccade. We make saccades when we move our eyes from one stationary object, such
as a sleeping cat, to another, such as a clock on the wall. We also make saccades as we
read, moving our fixation from one word to the next. If an object abruptly appears at
our periphery, activity in the area of the superior colliculus responsible for that area in
visual space becomes active, leading to a saccade, that is, a sudden movement directly
from our fixation to the new point in space. Think of a light going on in an adjacent
room. The light surprises you, and you make a sudden eye movement toward the
source of the light. Saccades are also the responsibility of the superior colliculus.
Another interesting feature of the superior colliculus is that it receives input from
other sensory systems, most noticeably the auditory system and the somatosensory
system. This allows our eyes to be directed quickly to the location of a sound or a
touch. Thus, when someone taps our shoulder, we quickly move our eyes in that
direction. This is adaptive—we want to be able to orient our visual system to any
surprising stimulus, regardless of how it is detected. We feel a strange tug on our
arm, and we want to be able to look at what is happening. We hear a loud sound
off to our side, and we want to be able to see what caused the noise. Interestingly,
these responses in the superior colliculus are additive. If we visually detect something
at the periphery and hear a loud sound off to the same side, our superior collicular
response will be larger than for either stimulus alone. In this way, the superior collic-
ulus is thought to be an area of multisensory integration (Stein & Meredith, 1993).
TEST YOUR KNOWLEDGE
1. What are the differences between saccades and smooth-pursuit eye movements?
2. How does the superior colliculus control eye movements?
THE PRIMARY VISUAL CORTEX
4.3 Explain the nature of the retinotopic organization of V1 and the organization of V2.
We now return to the main visual pathway. After leaving the LGN, the next synapse
in the visual pathway is in the primary visual cortex, in the occipital lobe of the brain.
The primary visual cortex, unfortunately, is referred to by a number of names. In dif-
ferent contexts, you may hear different nomenclature, but all refer to the same area of
the brain. So get ready for a list of terms that all refer to the same area of the brain:
Superior colliculus: a
structure located at the top of
the brain stem, just beneath
the thalamus, whose main
function in mammals (including
humans) is the control of eye
movements
Smooth-pursuit eye
movements: voluntary
tracking eye movements
Saccades: the most common
and rapid of eye movements;
sudden eye movements that are
used to look from one object
to another

100 Sensation and Perception
primary visual cortex, V1, striate cortex, area 17, and BA 17. Here’s the significance
of each term. The terms primary visual cortex and V1 refer to its position along the
flow of information in the visual system—the first area in the cortex to receive visual
information. Striate cortex refers to the way the brain cells in this area look under
certain staining conditions. They appear striated, whereas other adjacent areas of the
occipital cortex are “extrastriate.” Area 17 and BA 17 refer to the area’s position on
Brodmann charts. We use the terms primary visual cortex and V1 in this text, but
it is important to be aware of the other terms, as they are used just as frequently in
other discussions of the human visual system. When V1 is the specific topic being
discussed, we will use the term V1 for brevity. When the primary visual cortex is sec-
ondary to the main discussion, we will switch to the term primary visual cortex to
avoid excessive jargon.
The cerebral cortex is the outer surface of the brain. Indeed, the word cortex
comes from a Latin term for a tree’s bark. The cerebral cortex is only 2 mm thick,
but it is essential for all higher perception and cognition in humans and other mam-
mals. The cerebral cortex is divided into four lobes (Figure 4.7). The frontal lobe,
located behind the forehead, is the seat of higher cognition—thinking, planning, and
speaking, as well as a variety of motor functions. The temporal lobe is responsible
for memory, language comprehension, and auditory perception. The parietal lobe is
involved in attention and somatosensory perception. The occipital lobe, at the back
of the brain, is the visual cortex. This large area is responsible for vision and nothing
else. At the very back of the visual cortex is the area known as V1.
Primary visual cortex (V1):
the area of the cerebral cortex
that receives input from the
lateral geniculate nucleus,
located in the occipital lobe
and responsible for early visual
processing
Retinotopic map: a point-
by-point relation between the
retina and V1
Cortical magnification: the
allocation of more space in
the cortex to some sensory
receptors than to others; the
fovea has a larger cortical area
than the periphery
FIGURE 4.7 The
Four Lobes of the Brain
There are four lobes of the
brain: frontal, temporal,
parietal, and occipital. The
occipital lobe is the visual
area of the brain. This
figure shows the location
of V1 within the occipital
lobe. V1 is the first area in
the cortex that receives
input from the retinae.
Parietal lobe
Occipital lobe
V1
V4
MT
V2
V3
Frontal
lobe
Lateral
fissure
Temporal lobe
Central sulcus

101 Chapter 4: Visual System: The Brain
You can easily locate your own V1. Run your hand
along the back of your head, just above the neck. At the
very back of your head is a slight bump, known as the
inion, on the occipital bone. Press your finger against that
bone. Underneath the inion is V1. Because your inion is
right at the center or your head, your finger will be above
both the left and right V1s. You can also see its position
in Figure 4.8.
Mapping the Eye on the Brain
One of the important characteristics of the visual cortex is
that it is highly organized. This means that its anatomical
structure can be correlated directly with its function. That
is, if we know where a neuron is, we know what it does for
the visual system. This knowledge has come about from
more than 50 years of intensive study, starting with the
work of Hubel and Wiesel (see Hubel & Wiesel, 2005). We
now take a look at this exquisite organization.
The primary visual cortex is a bilateral structure—
there is one in the left hemisphere that receives input
from the right visual world, and there is one in the
right hemisphere that receives input from the left visual
world. The left and right V1s meet in the middle of the
brain (Figure 4.9). The area where they meet in the cen-
ter is the area responsible for representing the fovea, or
the area in space we are looking at.
The left V1 receives input from the left LGN, and the
right V1 receives input from the right LGN. As we saw
above, this means that each V1 will receive input from
the opposite half of the visual world. The left V1 receives
input from the left half of each retina, which corresponds
to the right half of the world. More than that, adjacent
locations on the retina project to adjacent points on V1.
Think of the retina as a terrain and the V1 as a map of
that terrain. There is a point-by-point relation between
the retina and the V1, as seen on a topographic map. As a
result, V1 is said to have a retinotopic map of the retina
(see Figure 4.9).
Just as maps of the world are not exactly the same as
the original terrain, neither is V1 a precisely exact map
of the retina. Some regions of the retina get to take up a
much greater proportion of V1 than others. This feature
is called cortical magnification. Cortical magnification
means that there is more space in the cortex devoted to
some sensory receptors than to others. In this case, the
fovea has a larger cortical area than the periphery. In par-
ticular, the fovea takes up a huge proportion of V1 and
the rest of the visual cortex. Indeed, the fovea is less than
1% of the retina in terms of size, but its representation
takes up over 50% of V1 (Mason & Kandel, 1991). That
©
S
hu
tt
er
st
oc
k.
co
m
/A
M
A
FIGURE 4.8
The Inion Marks the
Occipital Lobe
On this bald person’s skull,
you can see a prominent inion.
Right underneath that bone at
the back of the skull is V1 in
the occipital cortex.
FIGURE 4.9
Organization of the Visual World in V1
As in earlier areas in visual processing, the right V1 represents
the left visual world, and the left V1 represents the right visual
world. Note that the representation of the point of fixation, or
the foveae of the retinae, is toward the middle of V1, with the left
and right V1s adjacent to each other.
1
Left eye Right eye
Optic chiasm
Optic tract
Optic nerve
Foveal image
Superior
colliculus
Optic
radiations
Visual cortex
2 3 4 5 6 7 8 9
12345
1
1
2
2
3
3
4
4
5
5
5
5
6
6
7
7
7
6
8
8
8
23456789
9
9

102 Sensation and Perception
is a lot of the brain’s real estate in the cortex dedicated to a small portion of the
retina. Of course, it is the fovea that is specialized for color vision and visual acuity,
and these features are often the important features we need to determine what an
object is. So it is not surprising that representation of the fovea of the retina takes up
so much space in the cortex.
V1 is a six-layered structure, like all of the cerebral cortex. Each layer is num-
bered 1 through 6. However, in V1, most anatomists divide layer 4 into at least
three separate sublevels, and one of these sublevels is typically also divided into still
smaller sublevels. Layer 4 is the critical layer that receives input from the LGN. As we
discussed earlier, there are three distinct pathways coming from the retina through
the LGN to V1. Two of these arrive at V1 in layer 4, the magnocellular synapses
in sublayer 4Cα and the parvocellular layer synapses in sublayer 4Cβ (Yabuta &
Callaway, 1988) (Figure 4.10).
Receptive Fields of V1 Cells
When Hubel and Wiesel set out to do research on the anatomy and physiology of
V1, they expected to find receptive fields in V1 similar to those that were found in
retinal ganglion cells and in the LGN—that is, the basic center-surround organiza-
tion. What they found was much more complicated. They found a host of different
types of cells that had differing sensitivities to objects in their receptive fields. For
example, they found that V1 cells in cats were most sensitive to bars of different
orientations in different locations in the visual world. Thus, a particular neuron in
V1 might respond maximally to a vertical bar at 3 degrees off the fovea on the left.
An adjacent neuron in V1 might respond maximally to a bar 2 degrees off vertical
bar at 3 degrees off the fovea to the left. And this pattern would continue from one
cell to the next. They had not expected the organization to be so logical, but cells
seem to line up in columns sensitive to the orientation of objects in particular areas
in the visual field.
They also discovered two distinct kinds of cells in V1, which they called simple
and complex cells. We turn to them next.
FIGURE 4.10 V1 Layers
This illustration shows the multiple layers of V1. Layer 4 has been subdivided into three sublayers (A, B, C). Layer 4C is
then further subdivided.
O
ut
er
s
ur
fa
ce
o
f c
or
te
x
in
a
re
a
V
1
1 2 3
4
A B C 5
α β
6
White
matter
Area V1
Simple cells: V1 neurons that
respond to stimuli with particular
orientations to objects within
their receptive field; the preferred
orientation of a simple cell is
the stimulus orientation that
produces the strongest response
Orienting tuning curve: a
graph that demonstrates the
typical response of a simple cell
to stimuli or different orientations

103 Chapter 4: Visual System: The Brain
ISLE 4.2
Simple Cells
FIGURE 4.11
A Bar Detector
This cell’s receptive field allows
it to respond to a bar of light
at a particular orientation. It is
best stimulated by a light bar
surrounded by a dark field.
+
+
+
+








Simple Cells
Simple cells are V1 neurons that respond to stimuli with particular orientations to
objects within their receptive field. Simple cells are found in layer 4B of V1 and receive
input primarily from layer 4C of V1. Like cells in the LGN, they have clear excit-
atory and inhibitory regions. But unlike LGN cells, they have orientation selectivity
rather than center-surround visual fields. Hubel and Wiesel (1959) found that elongated
stimuli that looked like bars seemed to be particularly effective stimuli for these cells.
Indeed, they found that some cells wanted dark bars on a light background, and others
responded to white bars on a dark background. The bars may also occur at varying
angles of orientation, and these vary in a predictable pattern (Figure 4.11). This selec-
tive firing rate to the orientation shows the selectivity of the cell to orientation. You can
find an interactive illustration of simple cells on ISLE 4.2.
The preferred orientation of a simple cell is the stimulus orientation that produces
the strongest response from the simple cell. Experiments using single-cell recordings
demonstrate orienting tuning curves for any particular simple cell in V1. These ori-
enting tuning curves are graphs that demonstrate the typical response of a simple cell
to stimuli of different orientations. Such a curve can be seen in Figure 4.12. As you
can see in the curves in Figure 4.12, simple cells respond best to a stimulus with a
particular orientation; as the orientation gets larger or smaller, the response of the
cell decreases. If the orientation is greatly off, the cell will not respond at all. Other
cells have different peak sensitivities. Thus, V1 indicates the orientation of lines in
the visual world by having select cells respond to different angles of orientation.
Complex Cells and V1 Responses to Visual Features
Complex cells are also neurons in V1 that respond optimally to stimuli with particular
orientations. But, unlike simple cells, they respond to a variety of stimuli across different
locations. For example, a complex cell will respond to a dark bar on a light background
and a light bar on a dark background. In contrast, a simple cell responds to one but not
the other. Moreover, complex cells do not have peak location sensitivity, as simple cells
do. That is, they will respond equally well to an optimal orientation regardless of where
Complex cells: neurons in V1
that respond optimally to stimuli
with particular orientations;
unlike simple cells, they respond
to a variety of stimuli across
different locations, particularly to
moving stimuli
FIGURE 4.12 Tuning
Curves of Simple Cells in V1
You can see here the response
patterns of two simple cells. One
is tuned best to a perpendicular
90-degree bar, whereas the other
one is tuned to a 60-degree bar.
You can see that the cells will
respond to other orientations, but
they peak for one orientation.
60
50
40
Orientation
tuning curve of
Simple Cell B
R
e
sp
o
n
se
a
b
o
v
e
b
a
se
li
n
e
(s
p
ik
e
s/
se
c)
Orientation
tuning curve of
Simple Cell A
Responses to bar
oriented at 90°
30
20
10
0
0° 30° 60° 90° 120° 150° 180°

104 Sensation and Perception
it is within their receptive field. Some complex cells also receive
input from both eyes and may be involved in depth perception
(Read, 2005). Complex cells also respond best to moving stim-
uli. For moving stimuli, some complex cells are responsive to
movement in one direction, whereas other complex cells are
responsible for movement in the other direction. Complex cells
are found in layers 2, 3, 5, and 6 of V1, but not layer 4. You
can find an interactive illustration of complex cells in ISLE 4.3.
We have focused on the observation that simple and com-
plex cells are particularly sensitive to edges and bars with
specific orientations. But it is also important to keep in mind
that neurons in V1 are tuned to many different features,
including color, motion, depth, direction, length, and size.
This is particularly true of the complex cells. There are also
neurons in V1 called end-stopped neurons. End-stopped
neurons respond to stimuli that end within the cell’s receptive
field. If the pattern continues beyond the receptive field, these
cells do not respond as greatly. Because of this pattern, end-
stopped cells are considered to be involved in the detection of
corners and the boundaries of shapes.
The Organization of V1
When Hubel and Wiesel were recording from single cells, they
recorded from several cells in a single penetration of the elec-
trode into the brain. They recorded from one cell, and then they moved the electrode
slowly further into the brain until they found another cell (Figure 4.13). Then Hubel
and Wiesel would record from that cell and determine its receptive field, and repeat the
procedure again. If their electrode entered the brain perfectly perpendicular to the brain’s
surface, they found an interesting pattern. All of the cells in the perpendicular column
responded to a bar in the same location on the retina, thus demonstrating the same recep-
tive field (see Hubel & Wiesel, 1979). This is consistent with the idea of a retinotopic map.
Hubel and Wiesel also discovered that some cells prefer to respond to inputs from
one eye, and other cells prefer to respond to the other eye. This is called the ocular domi-
nance of the cell. These were also organized in columns, with some cells responding more
to stimuli from the right eye and some to stimuli from the left eye (Figure 4.14).
Moreover, Hubel and Wiesel also found that simple cells all selected for bars at par-
ticular locations. If they inserted their electrode at a different angle, they could find cells
that all responded to the same location, but for different orientations. At a different
angle, the electrode might find cells that responded to the same orientation along an
adjacent column. Thus, this vertical arrangement of cells that all responded to cells in
the same orientation in the same retinal location Hubel and Wiesel called a column.
Next, they noticed that adjacent columns responded to lines that were tilted only slightly
differently from one another. In fact, the columns formed an organized pattern accord-
ing to their orientation. It was later discovered that orientational selectivity goes in one
direction in V1, whereas the input from the two eyes goes in the other direction in V1.
Thus, Hubel and Wiesel found both ocular dominance columns and orientation col-
umns. The ocular dominance columns are made up of neurons that respond only to
one eye, and these columns are perpendicular to orientation columns, which selectively
respond to small variations in orientation. Ocular dominance columns alternate systemati-
cally between left-eye and right-eye dominance. Orientation columns change systematically
across orientations (Hubel & Wiesel, 1962). When ocular dominance columns and orienta-
tion columns are combined, they form something Hubel and Wiesel called a hypercolumn.
ISLE 4.3
Complex Cells
End-stopped neurons:
neurons that respond to stimuli
that end within the cell’s
receptive field
Ocular dominance column:
a column within V1 that is made
up of neurons that receive input
from only the left eye or only the
right eye
Orientation column: a column
within V1 that is made up of
neurons with similar responses
to the orientation of a shape
presented to those neurons
FIGURE 4.13 Orientation in V1
When an electrode is inserted into a column of cells in V1, all
respond to a specific orientation of bars.
Microelectrode
1
2
3
4A
2 mm
4B
4C
5
6
Outer surface
of cortex
Layer of cortex

105 Chapter 4: Visual System: The Brain
A hypercolumn is a 1-mm block of V1 contain-
ing both the ocular dominance and orientation
columns for a particular region in visual space.
This view of V1 persisted into the 1980s,
when researchers discovered that there was more
to the story. Hold on to your hat here, because
we are about to throw a whole bunch of new
terms at you. Livingstone and Hubel (1984)
discovered blobs and interblobs interspersed
within V1 (also see Wong-Riley, 1979). Blobs
are areas within V1 sensitive to color, whereas
interblobs are areas sensitive to the orientation
of an object. Recall that in the LGN, there are
two types of cells, the magnocellular and the
parvocellular. These two layers divide into three
types of cells in the cortex: the blobs, the inter-
blobs, and the cells in layer 4B. The interblob
cells respond as the simple cells described ear-
lier. The blobs show color responses, and the
layer 4B cells respond well to moving stimuli
and stimuli of very low contrast.
Now we need to integrate the blobs, interblobs, and layer 4B into the organization
of the striate cortex that has been discussed. This organization is seen in the hypercol-
umn. The orientational selectivity goes in one direction, and the ocular dominance goes
at right angles. In each set of orientation columns for each eye, there are two blobs.
Layer 4B runs throughout all of the cortex and cuts across all of the columns. This
hypercolumn is one functional unit, processing all of the information from one region
of the cortex. Adjacent hypercolumns process information from adjacent regions of
the cortex, and it is these hypercolumns that make up the topographical map discussed
earlier. The portions of the hypercolumns that are active at any point in time will indi-
cate the features of what is stimulating that region of the retina. Figure 4.15 gives you
Hypercolumn: a 1-mm block
of V1 containing both the ocular
dominance and orientation
columns for a particular region
in visual space
Blobs: groups of neurons within
V1 that are sensitive to color
Interblobs: groups of neurons
that are sensitive to orientation
in vision
FIGURE 4.14 Ocular
Dominance Columns in V1
Alternating columns in V1
receive signals from either the
left eye or the right eye through
the LGN.
1
2
3
4
5
6
C
or
tic
al
la
ye
rs
Left-eye
ocular dominance slab
Right-eye
ocular dominance slab
Blobs
Color code for orientation columns
FIGURE 4.15 Hypercolumns in V1
The hypercolumns organize the orientation columns, the location columns,
and the ocular dominance columns into a predictable pattern. The image
here is a schematic showing everything at right angles. In real mammalian
brains, the pattern may be curved or folded in on itself. In this figure, you can
also see the “blobs,” which are designed to detect color.
Blobs
Right-eye
ocular
dominance
Left-eye
ocular
dominance
Column
of cells
Row of cells with different orientational preferences
1 mm

106 Sensation and Perception
a sense of what all this would look like in the brain if everything were all straightened
out. For an interactive illustration of hypercolumns in V1, examine ISLE 4.4.
We understand that we are advancing a very complicated picture of the brain.
Many researchers have spent their careers investigating this so that we can under-
stand the visual system. But in other ways, it is remarkable that nature could have
designed such an incredible mechanism for decoding visual information and that it
does so in such a systematic fashion. We think that the organization of V1 is one of
the great accomplishments of nature and the work to understand it one of the great
feats of science.
TEST YOUR KNOWLEDGE
1. What is a hypercolumn? What are the subcomponents of it? How were these columns
discovered by neuroscientists?
2. Describe the relation of retinotopic organization to representation in the external world.
V2 AND BEYOND
You might wonder how the visual system goes from detecting edges and bars at partic-
ular angles to perceiving waving palm trees, skittering dragonflies, and smiling people.
This question has also motivated researchers for some time. Although this question
cannot be adequately answered yet, some progress has been made. To begin to answer
this question, V1 is not the end of visual processing. Indeed, it is more the beginning.
Information leaves V1 and proceeds to a
number of other areas within the occipital
lobe. Many of these areas eventually project
to sites in the temporal lobe or to sites in
the parietal lobe. The number of such dis-
tinct regions identified in the occipital lobe
grows as more research continues, but we
will touch on a few of the important areas
and also describe the two critical pathways
that emanate from V1.
V2
After information leaves V1, it travels to
other areas in the occipital cortex. These
areas are collectively called the extrastri-
ate cortex or secondary visual cortex.
The visual signals that leave V1 go in many
different directions. One of the major
pathways is from V1 to the adjacent V2
region (Figure 4.16). There are three dis-
tinct regions within V2, which match up directly with the three different types of cells
in V1. The blobs connect to the thin stripes. Layer 4B connects to the thick stripes,
and the interblobs connect to the interstripes (Zeki, 1993). It is possible to deduce
something of how these regions of V2 respond on the basis of their inputs from V1.
Thin stripes will have color responses, thick stripes will be sensitive to motion, and
interstripes will be sensitive to shape and position. V2 cells also combine input from
both eyes (Pasternak, Bisley, & Calkins, 2003).
ISLE 4.4
Hypercolumns
FIGURE 4.16 Visual Areas of the Rhesus Macaque Brain
This illustration shows a rhesus macaque brain. These monkeys have occipital
lobes very similar to those of humans. You can see the relative locations of V2
and V1, as well as other areas of the occipital cortex.
V1
V2
V3
V3A
V4
V5
V5A
Extrastriate cortex
(secondary visual cortex): the
collective term for visual areas
in the occipital lobe other than V1

107 Chapter 4: Visual System: The Brain
Other regions, such as V3, V4 (color processing), and V5 (also known as MT, a
motion detection area), receive input from only some of the cell types in V1 and V2.
V1 and V2 send information in parallel to these regions. We return to these higher
areas of visual processing later, after we discuss the dor-
sal and ventral pathways. V2, however, still seems to be
involved in representing what is out there rather than
making sense of it.
TEST YOUR KNOWLEDGE
1. What is meant by the term extrastriate cortex? How
does it differ from the striate cortex?
2. What are the purported functions of areas V3, V4,
and V5 (MT)? What might happen if these areas are
damaged?
FUNCTIONAL PATHWAYS
IN THE VISUAL CORTEX
4.4 Compare the difference between the dorsal
pathway and the ventral pathway.
We know the final goal of the visual system because most
of us, except for those with extreme visual impairment,
know the end result: fluent perception of the visual world
around us. Because we mostly take the act of seeing for
granted, we seldom consider how complex visual percep-
tion is. We have seen the complexity of the eye and the intricacy of visual processing
in V1. We next consider two essentially parallel pathways that have been discovered
within the visual system. We have begun to discuss these pathways, as we have already
introduced the terms P pathway (and K pathway) and M pathway as they relate to the
retinal ganglion cells. However, in V1, these pathways show very distinct organization,
and even more so, as these pathways leave V1 toward other regions of the cortex. The
P pathway has been described as the “what” or ventral pathway and the M pathway
as the “where” or dorsal pathway (DeYoe & Van Essen, 1988; Mishkin, Ungerleider,
& Macko, 1983). You can see the anatomy of this system in Figure 4.17. Let us take a
closer look at each of these systems.
The most straightforward way of learning these pathways is to examine Figure
4.18. This figure shows the flow of information from the retinal ganglion cells through
the LGN and occipital lobe and into other areas of the brain, including the temporal
and parietal lobes. What is striking about these pathways is that (a) very early in the
visual system, information is being sorted and channeled into different directions, and
(b) these pathways are both anatomically distinct and functionally separate. The ven-
tral pathway starts with bistratified and midget retinal ganglion cells. These retinal
ganglion cells connect with the koniocellular and parvocellular layers of the LGN,
respectively, which in turn project to V1 layers 2 and 3 and sublayer 4Cβ, respectively.
From V1, the signal is sent to other visual areas in the occipital cortex, which we
discuss shortly, and then to an area in the temporal lobe known as the inferotempo-
ral cortex. This area is known from a great deal of research to be involved in object
V2: the second area in the
visual cortex that receives input;
often considered the area that
starts with visual associations
rather than processing the input
(sometimes called the prestriate
cortex)
Ventral pathway: starts with
midget and bistratified retinal
ganglion cells and continues
through the visual cortex into
the inferotemporal cortex in the
temporal lobe; often called the
“what” pathway, as it codes for
object identification as well as
color vision
Dorsal pathway: starts with
parasol retinal ganglion cells
and continues through the
visual cortex into the parietal
lobe; often called the “where”
pathway, as it codes for the
locations of objects and their
movement
FIGURE 4.17
The Ventral and Dorsal Pathways in the Brain
This illustration shows the important visual areas of the occipital lobe and
the flow of information in both ventral and dorsal pathways. The ventral
pathway flows through V2 and eventually into the temporal lobe and is
concerned with object identification, as well as color perception. The
dorsal pathway flows through V2 to area MT and then to the parietal lobe.
It is concerned with where objects are in visual space as well as motion.
Posterior
parietal cortex
Ventral stream
Dorsal stream
V1
V4
MT
V2
V3
Inferotemporal cortex

108 Sensation and Perception
FIGURE 4.18 The Flow of Information in the Brain
This figure illustrates the flow of information schematically from retinal ganglion cells to higher areas in the cortex of the human
brain. Note that the ventral and dorsal pathways are defined very early, starting in the retinal ganglion cells. Thus, from early in visual
processing, the “what” and the “where” pathways are distinct.
Retina
Parasol
Midget
Bistratified
Retinal
Ganglion
Cells
(RGCs)
Lateral
geniculate
nucleus
(LGN)
Layers of V1 V2
Cortex
Magnocellular
layers
Parvocellular
layers
Koniocellular
layers
12/34A4B4Cα4Cβ56
Blobs
Blobs
Blobs
Thick bands
(motion)
Pale bands
(form)
Thin bands
(color)
Parietal
Cortex
(perceiving
space and
motion)
V4
(form,
color)
Inferotemporal
Cortex
(object
recognition)
V3
V3
V5/MT
(motion)
V4/V5
Interblob
region
Dorsal (“where”/“how”) pathway
Ventral (“what”) pat
hway
recognition. Hence, the ventral system is often called the “what” system. The dorsal
system starts with the parasol retinal ganglion cells, which project to the magnocellular
cells in the LGN. From there, the pathway leads to sublayer 4Cα in V1. The pathway
then projects to V2 and from there to such areas as MT, ending up in the parietal
cortex. This pathway codes for place and movement, hence its nickname, the “where”
system. Ultimately, what remains unclear is where in the brain the systems rejoin so
that we see a unified perception of the world.
We consider an experiment that was done to establish the reality of the dorsal and
ventral pathways. Keep in mind that all the knowledge we have been presenting is the
result of careful experimentation, like the study we are about to describe. In this study,
Mishkin et al. (1983) trained rhesus monkeys (Macaca mulatta) to do two different
tasks. One task was called a landmark task, which required the monkeys to remember
the location of an event, and the other task was called an object task, which required the
monkeys to learn a particular object. In the study, monkeys saw two containers, one of
which contained food. In the landmark task, the monkey was required to select the con-
tainer that was closer to a specific landmark in the room. In the object task, the monkey
was required to select the container that was covered by a particular object. Rhesus mon-
keys learn these tasks quickly and then seldom make errors. Note that for each task, the
monkey is expected to do something that emphasizes a different system. In the landmark
task, the monkey is focusing on where an object is (the dorsal stream), whereas in the
object task, the monkey must remember what the object is (the ventral stream).
Mishkin et al. (1983) then created damaged areas in parts of the monkeys’ brains
(Figure 4.19), a process known as lesioning. In some monkeys, they lesioned the infero-
temporal cortex, an important area in the ventral or “what” pathway. In other monkeys,
they lesioned the parietal lobe, part of the dorsal or “where” pathway. After recovering
from surgery, the monkeys were again presented with the landmark task and the object
task. The monkeys with damage to the inferotemporal cortex showed normal perfor-
mance on the landmark task but were impaired on the object task, consistent with the
contention that the inferotemporal cortex is involved in establishing the “what” of an
object. In contrast, the monkeys with parietal lobe lesions showed normal performance

109 Chapter 4: Visual System: The Brain
Lesion in
inferotemporal
cortex
Lesion in
parietal cortex
Bin with food
Landmark
Landmark task
(a “where” task)
Find food in bin closer to landmark.
Object task
(a “what” task)
Find food in bin under square object.
Square object
FIGURE 4.19 A Lesion
Study Illustrating the Reality
of the Dorsal and Ventral
Pathways
Monkeys learned one of two
tasks, a landmark task and
an object task. Lesions to the
parietal lobe interfered with the
landmark task, and lesions to the
inferotemporal lobe interfered
with the object task. This study
provides support for the view
that the dorsal and ventral
streams are separate.
on the object task but were impaired on the landmark task, consistent with the conten-
tion that the parietal lobe is involved in establishing where an object is. Thus, damaging
the inferotemporal cortex interferes with visual object recognition, but damaging the
parietal lobe interferes with seeing the spatial relation of objects.
Although it is unethical to do such an experiment with human beings (and many would
argue the experiment with rhesus monkeys was unethical as well), accidents have resulted
in damage that approximates what Mishkin et al. (1983) did to the monkeys. In one case,
a patient known as D.F. suffered carbon monoxide poisoning, which damaged her ven-
tral pathway (Goodale, Milner, Jakobsen, & Carey, 1991). As a consequence of the brain
damage, D.F. showed severe deficits in her ability to perceive and therefore name objects
(a condition known as object agnosia). Despite her inability to perceive objects, which is
a function of her ventral system, she was able to grasp and manipulate them, presumably
because of her intact dorsal system. For example, when asked to estimate the length of an
object, her accuracy was severely impaired relative to a normal person. However, when
asked to lift the object, her hands adjusted appropriately to the size of the object.
The Ventral Pathway
The ventral pathway starts with midget and bistratified retinal ganglion cells and con-
tinues through the visual cortex into the inferotemporal cortex in the temporal lobe.
The ventral pathway is often called the “what” pathway, as it codes for object iden-
tification as well as color vision. After being processed in layers 2, 3, and 4Cβ of V1,
information is sent to V2 and from there to areas such as V3 and V4 in the extrastriate
cortex. V4 then sends information to the inferotemporal cortex. V4 has some interest-
ing properties. V4 neurons are sensitive to binocular disparity, helpful in recognizing
three-dimensional objects in space. Thus, V4 is seen as primarily involved in shape rec-
ognition (Pasupathy & Connor, 2002). However, V4 also seems critical in color vision:
Patients with damage to V4 have disorders of color vision (Zeki, 1993). We return to
V4 and disorders of color vision in Chapter 6.
From V4, information flows out the occipital lobe and into the temporal lobe
to an area of the cortex known as the inferotemporal cortex. The inferotemporal
cortex is the region in the temporal lobe that receives input from the ventral visual
pathway. One of its major functions is object identification. Early studies of the
inferotemporal cortex showed that its neurons are sensitive to highly specific kinds
of shapes (Bruce, Desimone, & Gross, 1981). Although some neurons in the infero-
temporal cortex are sensitive to basic visual features such as size, shape, color, and
orientation, the studies that attracted interest found neurons with specific responses
Object agnosia: an acquired
deficit in identifying and
recognizing objects even though
vision remains intact
Inferotemporal cortex: the
region in the temporal lobe that
receives input from the ventral
visual pathway; one of its
functions is object identification

110 Sensation and Perception
to such complex features as hands, paws, and faces. Indeed, research demonstrated a
region within the inferotemporal cortex called the fusiform face area (FFA). The FFA
seems to be a region specifically dedicated to recognizing familiar faces (Kanwisher
& Yovel, 2006). Single-cell recording studies with monkeys show that neurons
in the FFA are maximally responsive to the faces of other same-species monkeys.
Neuropsychological studies with brain-damaged patients show that patients with
damage to the FFA have deficits in identifying familiar faces (Schwartz, 2017; Susilo
& Duchaine, 2013). We return to the topic of face recognition in Chapter 5. Other
areas of the inferotemporal cortex appear to be sensitive to other kinds of object
recognition. Indeed, one study showed that a particular area of the inferotemporal
cortex is sensitive to the perception of individual persons. Cells in this area become
active when particular individuals are presented—either pictures of them or the
spelling of their names (Quiroga, Reddy, Kreiman, Koch, & Fried, 2005). Thus, the
ventral or “what” pathway is the neural pathway that allows us to perceive and
recognize the objects in our environment, regardless of whether they are the faces of
our family members or the books, pillows, clocks, smartphones, and eyeglasses that
populate our visual world.
The Dorsal Pathway
The dorsal pathway starts with parasol retinal ganglion cells and continues through
the visual cortex into the parietal lobe. The dorsal pathway is often called the “where”
pathway, as it codes for the locations of objects and their movement. After informa-
tion leaves V2 in the dorsal pathway, it is sent to an area in the occipital lobe known
as MT (for middle temporal, its location in the occipital lobe). MT is also known as
V5. MT is also connected to noncortical visual movement areas, such as the superior
colliculus (Born & Bradley, 2005). Single-cell studies with monkeys show that neu-
rons within MT are sensitive to the direction and speed of motion (Albright, 1984).
These data have been confirmed in human studies using functional magnetic resonance
imaging (fMRI) (Born & Bradley, 2005). Transcranial magnetic stimulation studies,
which render an area of the brain temporarily disrupted, have also been used to probe
human MT function. These studies show that interfering with MT functioning results
in a disruption of motion perception (Schenk, Ellison, Rice, & Milner, 2005). Finally,
patients with damage to MT have difficulties perceiving motion. We return to these
issues in Chapter 8.
After the signal leaves MT, it is sent across the occipitoparietal junction into the
parietal lobe. Specific regions in the parietal lobe that are part of the dorsal pathway
include the anterior intraparietal, the lateral intraparietal, and the medial intrapari-
etal areas. These areas are involved in the visual guidance of action. Research in both
humans and monkeys suggests that these areas are involved in the visual guidance
of reaching for and grasping objects (Culham, Cavina-Pratesi, & Singhal, 2006). We
return to these topics in Chapter 8.
TEST YOUR KNOWLEDGE
1. What are the functions of the dorsal and ventral pathways?
2. Why might color information be important for the ventral pathway but not the dorsal
pathway?
MT (V5): an area of the
occipital lobe in the dorsal
pathway, specific to motion
detection and perception

111 Chapter 4: Visual System: The Brain
WHERE DOES VISION COME TOGETHER?
So, we know that object perception comes through the ventral pathway and that move-
ment and object location come through the dorsal pathway. But what we perceptually
experience are moving objects—a unified world. Thus, as a result of our discussions of
brain regions and visual pathways, you might be wondering where all of vision comes
together, that is, what region of the brain is responsible for giving us a unified and
common perception of the world. After all, when we look at the world, we see talking
people, running cats, and airplanes moving across the sky, rather than the objects and
their movement separately, as might be inferred from our discussion of pathways. One
can simplify the question: Where in the brain do these pathways come together to give
us this common perception? At present, such a region has not been found, and many
researchers think it does not exist. In fact, nowhere in the brain has a single loca-
tion been found where all visual information converges (Dennett, 1991; Zeki, 1993).
Remember, V1 and V2 are the last areas to have all of the visual information, and that
is only for the central visual pathway (Zeki, 1993). It now appears that conscious vision
happens with simultaneous or approximately simultaneous activation across all visual
areas. Perhaps important to this neurologically are the feedback loops that exist in the
brain. At every level, there is return feedback to earlier levels in the pathways from
higher levels in the pathways. Forward and backward connections are thought to play
a role in synchronizing a response in the brain, so that our perceptual experiences are
of whole objects and not fragmented parts. Thus, our experience of vision is distributed
at least across the areas of the visual cortex. So one answer to why we experience the
world as a unified whole is that the constant feedback loops from higher to lower levels
integrate our perceptions.
TEST YOUR KNOWLEDGE
1. In your view, will the search for an area of the brain that unifies visual processes into a
common perception be successful? Why or why not?
DEVELOPMENT OF THE VISUAL SYSTEM
We have seen that the brain organization involved in vision is complex, highly struc-
tured, and supremely well adapted. However, regardless of whether we have been
discussing the visual systems of cats, monkeys, or people, we have so far been consid-
ering mature individuals. An important question pertains to how our visual systems
mature and develop. In this section, we take up the issue of development. Often
critical in discussions of development is whether a particular ability requires practice
and experience or whether it will develop innately, regardless of experience. In other
words, is the organization of our visual system hardwired? That is, is it set specifically
by our DNA, does it follow a set course of development, and is this unchangeable?
Or are environmental inputs necessary for the visual system to develop normally?
If so, how flexible is our visual system, depending on that environmental change?
As we will see shortly, there is much flexibility in the development of the visual sys-
tem. Certain patterns appear to be fixed, but much can be altered by the organism’s
early environment.

112 Sensation and Perception
Is it nature (our genetics) or nurture (our experiences) that is more important
in how our sensory systems operate? Texts usually present the two extreme posi-
tions and contrast them. At one extreme, there is the genetic position, that all that
is needed for our sensory systems, or whatever biological or psychological factor is
being discussed, is coded in our genes. Sensory operation is thus essentially “pre-
ordained.” On the other side is the tabula rasa position, that we are blank slates at
birth, and experience writes on these slates to determine who we are going to be. The
fact is, however, that genes and the environment interact in incredibly complex ways
to allow development to occur, and nearly all scientists agree that this interaction is
inevitable and complex. Indeed, it may not be possible in many cases to specify what
is innate and what is learned when it comes to something as complex as the mam-
malian visual system. Certainly the interaction position fits all the data and studies
we discuss here.
One landmark research study on the development of the visual brain comes
from work by Held and Hein (Held, 1965, 1980; Held & Hein, 1963). These
studies, which some of you may find distasteful, concerned the development of
the visual system in young cats. The researchers carefully showed how important
environmental input is to the development of the visual brain. Held and Hein
raised kittens in complete darkness for a few weeks. After the initial period of
darkness, the kittens were divided into two groups. Both groups were still kept
in darkness most of the day but now spent an hour a day on a “kitten carousel”
(Figure 4.20). The kittens in one group were free to
move themselves. As they moved, the visual environ-
ment moved around them. Thus, they could correlate
their movement with the movement around them.
The second group of kittens were kept in little bas-
kets and were moved by the movements of the other
kittens. Thus, their movement was passive—tied not
to their own motor systems but to those of the other
kittens.
Later, the kittens were tested on a variety of visual
tasks. In all cases, the active kittens performed better
than the passive kittens, despite their identical visual
experience. For instance, in a test of visually guided
paw placement, the active kittens were better able to
extend a paw to touch a point on a flat surface than
were the passive kittens. In another test, the active kit-
tens were more likely to avoid a “visual cliff,” that is,
an illusion that a sharp drop-off is present where they
are walking. In yet another test, the active kittens were
better at the visual pursuit of a moving object, which
is important for carnivorous cats, which must chase
fast-moving prey. Indeed, the active kittens showed no
differences from cats raised normally with full visual
experience. It was only the control-deprived passive
kittens that did poorly on the visual tests. These results
suggest that there is flexibility in the development of
the visual system, that early visual–motor experience can later influence perceptual
abilities (Held & Hein, 1963).
At the same time that Held and Hein were doing pioneering work with the kitten
carousel, Hubel and Wiesel were also examining developmental issues in the visual
FIGURE 4.20 The Kitten Carousel Experiment
The illustration shows two kittens, both reared in an all-dark
environment, seeing vertical stripes in the experimental setup.
The kitten on the left is active in its exploration, whereas the kitten
on the right is yoked to the first. Because of its lack of active
experience, the yoked kitten shows slowed visual development
(Held & Hein, 1963).

113 Chapter 4: Visual System: The Brain
brain (Wiesel & Hubel, 1965). Like Held and Hein, they also focused on deprivation
studies, that is, studies in which young organisms (usually cats) were deprived of
visual experience of one form or another. In Hubel and Wiesel’s research, the goal
was to examine the receptive fields of individual neurons rather than the cats’ behav-
ior in general. They found that, consistent with the “nature” view of the brain, some
organization already existed at birth in the cats’ visual system. Even with no visual
experience, there was some organizational structure in the cats’ visual cortices (Daw,
2009; Hubel & Wiesel, 2005).
Wiesel and Hubel (1963) also showed that there is degeneration in the cortex if an
organism is denied visual experience. For example, in one study, kittens were deprived
of visual experience in one eye (monocular deprivation). The other eye allowed the
kittens to interact normally with their environment. Later, Wiesel and Hubel looked
at the receptive fields of neurons in the kittens’ V1s. They found clear evidence that
monocular deprivation resulted in a reduction in the number of cells that responded to
the deprived eye and fewer cells that responded to both eyes. However, such changes
occurred only if the kittens were deprived of visual experience during the first 3 months
of life. Visual deprivation after 3 months had no effect on neural organization. Thus,
Wiesel and Hubel’s studies suggest that both innate organization and environmental
inputs are important to the developing visual system (Daw, 2009).
Conducting such a deprivation experiment with human infants would be highly
unethical. So experiments cannot be applied to young human visual systems. But
correlational research can be done, examining changes in the visual cortex as infants
begin to mature. In one such study, Huttenlocher and de Courten (1987) found that
during early infancy (2–4 months), there is rapid growth of synapses within V1 of
the cortex. This is followed by a loss of synapses later in childhood (8–11 years). This
suggests that early visual experience is necessary for establishing the parameters of
vision, but unnecessary connections will later be pared down. Thus, it is likely that
human visual development follows a similar trajectory in terms of some innate pat-
terns elaborated on by experience.
TEST YOUR KNOWLEDGE
1. Is our visual system hardwired, or do learning and change take place as the mammalian
eye and visual system mature?
2. Describe the “kitten carousel” experiment. What does it tell us about visual development?
4.5 Interpret the concepts of blindsight and conjugate gaze palsy.
EXPLORATION: Blindsight
Patient T.N., born in Burundi, Africa, but living in
Switzerland, is a doctor who worked for the World Health
Organization. In 2003, while in his mid-50s, T.N. had two
serious and successive strokes over a period of 5 weeks,
each stroke causing lesions on one side of his visual cor-
tex. The first stroke affected his left visual cortex, and the
second stroke, 5 weeks later, destroyed his right visual
cortex (de Gelder, 2010). Subsequent anatomical testing
showed that T.N.’s entire primary visual cortex (V1) was
compromised (de Gelder et al., 2008). As a consequence
of these strokes, T.N. is now completely blind. He reports
no visual experience, though his overall intelligence and
verbal abilities remain completely intact. Not relying solely
on these subjective reports, de Gelder and her team ran
a series of routine tests of vision. The tests turned up
nothing—T.N. showed no evidence of seeing in these

114 Sensation and Perception
standard tests. Moreover, structural MRI revealed exten-
sive structural damage to V1, and functional MRI revealed
no activity in this region of the brain when T.N. was shown
visual stimuli (de Gelder et al., 2008). However, because
the strokes affected his visual cortex, his eyes themselves
were undamaged.
What makes T.N.’s case fascinating is that despite his
complete perceptual blindness, he still makes visual
responses. Indeed, in 2008, de Gelder et al. staged a
stunning demonstration of this. T.N. was told that a
cluttered corridor was empty and that he would not
need his cane to walk down the corridor. Reluctantly,
T.N. agreed. As he walked down the corridor, he was fol-
lowed by another person (a prominent researcher in this
area) to ensure that he would not stumble. When you
watch the video, you can see that despite his total blind-
ness, T.N. avoided the obstacles placed in front of him.
He sidesteps a box and avoids a tripod. After completing
the trip down the hallway, T.N. does not know what he
has avoided or whether there was anything in his way at
all. In other words, he made visually guided movements
in the absence of conscious sight. This is a phenome-
non known as blindsight. Blindsight is the presence of
visual abilities even though a person experiences blind-
ness because of damage to V1. The patient is subjectively
blind but makes accurate visual responses.
To see T.N. walking down the corridor, you can watch
a video at the Scientific American website (http://www
.youtube.com/watch?v=ACkxe_5Ubq8, see also ISLE 4.5).
The researcher who followed him down the hallway
reported that T.N. did not make vocalizations for which he
might listen for slight echoes,
the way a bat or dolphin
uses echolocation. However,
they had no sound-monitor-
ing equipment, so this could
not be completely ruled out. Amazingly, in your textbook
authors’ opinion, the researchers failed to do a simple con-
trol—to have T.N. negotiate an obstacle course with a blind-
fold on. If blindsight were the explanation for his avoiding
the obstacles, then cutting off input to the retinae would
render him unable to negotiate the course. Indeed, they did
only one trial with T.N., and they did not perform appropri-
ate control investigations. Nonetheless, this demonstration
is now a well-cited example of the abilities of blindsight
patients (de Gelder, 2010). Luckily, other research has been
done with T.N. and on other patients with similar condi-
tions that have used the necessary experimental controls.
T.N.’s case is unique in that his entire V1 was compromised
(Figure 4.21). In other cases of blindsight, the patient has
damage in some areas of V1 but not all. The result is a sco-
toma, an area of partial or completely destroyed cells, result-
ing in a blind spot in a particular region of the visual field.
In these patients, vision may be normal in most regions of
the visual field, allowing them to see normally under most
circumstances. However, when these patients keep their eyes
still, there is a blind spot (the scotoma) where they cannot
see or cannot see normally. In a number of patients tested,
blindsight may exist within the scotoma region (Weiskrantz,
1996). Such patients have been yielding an understanding
and exploration of blindsight for many years, but T.N.’s case
is unique in that the scotoma consists of his entire visual field.
The first studies of blindsight go back to the 1970s. At the
time, Lawrence Weiskrantz, an English neurophysiologist,
was examining the visual responses of monkeys with lesioned
V1s. He found that monkeys were still able to make visually
guided responses even without this region. He then focused
his research on the superior colliculus, the part of the brain
described earlier that guides eye movements. Weiskrantz rea-
soned that intact colliculi allowed the monkeys to respond
to visual stimuli in the absence of V1 (see Weiskrantz, 1996).
Indeed, despite their missing V1 regions, these monkeys
appeared to be behaving normally and responding appro-
priately to visual stimuli. However, he knew from neuropsy-
chological research that humans with damaged V1 regions
had visual problems known as scotomas, that is, areas in the
visual field in which they could not see. The blind field in the
visual world corresponds to the damaged area in V1, which
is responsible for that region in space.
Then Weiskrantz discovered patient D.B., and the two
began a long collaborative investigation of D.B.’s blindsight
(Weiskrantz, Warrington, Sanders, & Marshall, 1974).
D.B. was a 34-year-old British man, working as a computer
programmer. He was married with young children. He was
also an amateur musician and an amateur rugby player.
He started developing headaches, and his doctor sent him
to a neurologist. The neurologist discovered a large tumor
located in the primary visual cortex of his right hemisphere.
To ensure his long-term health, much of D.B.’s right V1 was
removed surgically along with the tumor, leaving him with
a large left-field scotoma. Nonetheless, his vision was still
ISLE 4.5
Navigation in Blindsight
Blindsight: the presence of visual abilities even though a
person experiences blindness because of damage to V1
Scotoma: an area of partially or completely destroyed cells,
resulting in a blind spot in a particular region of the visual field

115 Chapter 4: Visual System: The Brain
fine in his right visual field, and D.B. was able to return
to his normal life. However, he agreed to continue having
his scotoma investigated by Weiskrantz and his colleagues
(Weiskrantz, 1986). Note that D.B., unlike T.N., still has
much of his V1 intact, meaning that he can still see much
of the world. His left V1, responsible for his right visual
field, was just fine. He could see normally in this region.
His scotoma (blind spot), however, extended through much
of his left visual field.
When shown objects in his intact right visual field, D.B.
was able to identify and describe them as any sighted per-
son would. However, when objects were presented in the
scotoma in his left visual field, D.B. reported not being able
to see them, nor could he identify them. To D.B., he sim-
ply could not see in this field, and like T.N., D.B. was sub-
jectively blind in the scotoma region. To test for blindsight,
Weiskrantz used a forced-choice procedure in which D.B.
had to guess what he saw. In this procedure, D.B. might feel
as if he had not seen anything, but he had to make a response
anyway. In one study, Weiskrantz presented either a square
shape or a diamond shape, and D.B. had to guess which
was presented. Even though the image was presented to his
scotoma and the only difference between the two images
was orientation, D.B. guessed at the correct shape at a rate
significantly higher than chance. In another study, D.B. could
discriminate between X and O at above-chance rates. In yet
another study, D.B. had to determine whether an image in
his scotoma was stationary or moving. Again, D.B. reported
no conscious seeing of the movement, but his guesses were
significantly greater than chance (Weiskrantz, 1986). Since
then, this phenomenon has been observed in many patients
and continues with D.B. today as well (Tamietto et al., 2010).
Having established that patients with blindsight (a) are
truly without visual experience in their scotoma fields,
and (b) are able to make visual responses, the term
blindsight may seem slightly less paradoxical. Starting
with Weiskrantz et al. (1974), the predominant view is
that blindsight is mediated by mechanisms in the supe-
rior colliculus that continue to get input from the retina
even when the pathways to the visual cortex have been
damaged. Here is the basis of Weiskrantz’s view.
Patients such as T.N. and D.B. have extensive damage to
the visual cortex, which creates the experience of blind-
ness. However, their retinae continue to project axons,
FIGURE 4.21 Damage to Blindsight Patient’s Brain
An MRI study of patient T.N.’s brain, showing the bilateral destruction of the V1 region of his brain. This damage means that his V1 does
not process visual information. As such, T.N. is blind. However, other areas of the brain, such as the superior colliculus, allow him to
make visual responses. From de Gelder et al. (2008).

116 Sensation and Perception
which reach the superior colliculus and other noncorti-
cal areas of the brain. In people without brain damage,
the superior colliculus is an area of the brain involved
in the rapid movement of the eyes and the head toward
the source of an incoming visual stimulus. These mech-
anisms appear to be intact in patients with blindsight,
as they may move their heads toward incoming stimuli
even though they cannot see the stimuli. Consider T.N.—
he cannot see the obstacles placed in front of him, but
as he moves toward them, he slowly but surely avoids
them. His actions are consistent with the explanation that
blindsight is caused by spared noncortical routes from the
retinae to the brain.
Given that there are several routes from the retinae to
noncortical areas of the brain, one might expect to find
other kinds of blindsight in addition to object avoidance
and forced-choice discrimination. Indeed, research with
T.N. demonstrates a form of emotional blindsight (Pegna,
Khateb, Lazeyras, & Seghier, 2005). When presented with
images of fearful faces, T.N. made cringing expressions
himself and showed activity in his amygdala (an area of the
brain associated with emotion), even though he could not
report what the stimuli were that he was seeing. Remember
that T.N. shows no activity in his V1 during these studies.
Thus, it is likely that the responses are mediated by the
superior colliculus route. Interestingly, some research now
demonstrates that the superior colliculus projects to higher
order areas in the visual cortex (Ptito & Leh, 2007). This
may account for T.N.’s emotional responses.
Over the years, blindsight has drawn significant interest
from philosophers. For them, the intriguing phenomenon
is that someone is blind but is making responses to visual
stimuli. This seeming paradox is resolved when we think
of conscious seeing as being a function of V1 in the cortex.
When V1 is damaged, we lose conscious vision. However,
intact retinae continue to project to noncortical areas,
which allow us to respond to visual stimuli, despite the lack
of visual seeing.
APPLICATION: Conjugate Gaze Palsy
Ophthalmology, though nearly impossible to spell, is a
major area of medicine. Healthy vision is an important
aspect of human life for most people. However, vision
can be impaired by things as humdrum and simple as a
misshapen lens to complex disorders of vision, such as
agnosia (in which people see and can describe accurately
what they see, but they do not extract meaning). A work-
ing ophthalmologist must, in addition to working with
patients who have diseases of the eye, work with patients
who have incurred damage to any number of regions of
the brain, damage that manifests as visual problems. In
this section, we describe the conjugate gaze palsies, a
series of disorders of vision, brought about by damage to
the oldest part of the brain, namely, the brain stem.
Human beings, like all mammals, have two eyes. In many
ways, we take this for granted. In Chapter 7, we will dis-
cuss how having two eyes allows humans the ability to
create a binocular image, which enhances our ability to
see at depth. Humans and many other mammals use bin-
ocular vision, which requires that both eyes be trained on
the same object at the same time. If one eye is looking far
to the left and the other to the right, the corresponding
images are not focused on the same object and not only is
depth perception lost, but double images may occur. Try
“crossing” your eyes, and you may be able to produce
such a double image. Thus, having two eyes requires our
visual systems to have mechanisms in place that are able
to keep our eyes coordinated. In most of us, these mecha-
nisms are functional, and we seldom worry about whether
our eyes are focused on the same thing. The supranuclear
ocular motor system, located in the brain stem, is a system
concerned with linking eye movements in each eye to each
other. The supranuclear ocular motor system connects to
a great many regions in the brain that are used in visual
processing, including the frontal eye fields. However, in
conjugate gaze palsy, brain damage to the brain stem
affects the ability of our eyes to train on the same object
(Goldberg & Trattler, 2015). Conjugate palsies are some-
times seen as complications in Parkinson’s disease, after
cardiac surgery (Eggenberger, 2014), and when the pineal
gland is compromised (Goldberg & Trattler, 2015).
Parinaud’s syndrome occurs from lesions in the superior
colliculus. Rather than irregular movements and difficul-
ties coordinating the two eyes, patients with Parinaud’s
Conjugate gaze palsies: neurological disorders that affect
the ability of the eyes to coordinate their movements; this
inability to move together may affect eye movements in both
vertical and horizontal directions

117 Chapter 4: Visual System: The Brain
syndrome have a paralysis of gaze. They must move their
heads, rather than their eyes, to track motion.
Conjugate gaze palsies are, therefore, neurological disorders
that affect the ability of the eyes to coordinate their movements.
This inability to move together may affect eye movements in
both vertical and horizontal directions (Goldberg & Trattler,
2015). The cause of these palsies lies not in the peripheral
musculature of the eyes but in locations in the brain, partic-
ularly in the brain stem region that regulates eye movements
(Xie, Yu, Wang, Liu, & Meng, 2016). Let us look at one form
of conjugate gaze palsy in more detail. Internuclear ophthal-
moplegia results from selective damage to the medial longi-
tudinal fasciculus (MLF). The MLF is a bilateral structure in
the brain stem, consisting of mostly axons (i.e., white-matter
structure). In this condition, if a patient is asked to look to the
left, the left eye will make the appropriate movement, but the
right eye will not move. Similarly, if a patient is asked to look
to the right, the right will make the appropriate movement,
but the left eye will not move. When this occurs, the patient
will have “double vision” and blurry vision because the eyes
are not focused on the same object.
When areas like the MLF and other related areas are
damaged, the result is often double vision (seeing two of
individual objects in the environment), blurry vision, and
the making of unwanted eye movements (Eggenberger,
2014). These symptoms can be quite troubling, as they
impair many basic functions, such as visual guided move-
ments, including walking and driving as well as reading.
For example, Eggenberger described a patient who was
given a strong antibiotic as treatment for an unrelated
eye problem. The antibiotic induced changes in the brain
stem, which left him with gaze palsy symptoms. The
patient, though otherwise healthy, found he could no lon-
ger drive and that reading was impaired.
There is no reliable treatment that works every time.
However, there are a variety of ways of helping peo-
ple with gaze palsies. In general, doctors will work
to remove the cause of the initial impairment by restor-
ing blood supply to the affected areas. Still, treatments
vary, as in some cases, the symptoms of gaze palsies
will dissipate with time, presumably as other adjacent
areas in the brain stem take up the regulation of coor-
dinating eye movements. However, for some patients,
the symptoms persist. In these cases, there are various
remedies, depending on the patient. In some cases, special
prisms can be constructed for eyeglasses to bring the eyes
back into common focus (Eggenberger, 2014). In other
cases, it may be possible to have the patient work on ver-
gence exercises, to strengthen the control over eye move-
ments (Eggenberger, 2014). In cases in which the damage
leads to tremor in addition to uncoordinated movements
of the eyes, drugs similar to those used in epilepsy may
be tried. These drugs may include antitremor drugs such
as gabapentin and antiseizure drugs such as clonazepam
(Aladdin, Scozzofava, Muayqil, & Saqqur, 2008). Because
these drugs reduce tremor, they may allow patients to
regain control over their eye movements. Because of the
underlying cause, it is difficult to predict the prognosis in
most cases of conjugate gaze palsy. Patients may recover
rapidly in some cases, but, in other cases, recovery does not
occur (Aladdin et al., 2008).
CHAPTER SUMMARY
4.1
Identify the anatomy of the optic chiasm and how
that affects the lateralization of vision in the brain.
When information from the optic nerve enters the optic chi-
asm, information crosses over, so that the axons from the
ganglion cells from the right half of the right retina and the
ganglion cells from the right half of the left retina combine,
forming the right optic tract, which then proceeds to the
right hemisphere of the brain (and similarly for the other half
of the system).
4.2
Diagram the anatomy of the lateral geniculate
nucleus and the superior colliculus, and describe
their roles in visual processing.
The optic nerve projects to the lateral geniculate nucleus
(LGN). The LGN is a six-layered structure in the thalamus
that serves as a relay point for the transmission of visual
information, although processing of information also occurs
in the LGN. Each of the three types of retinal ganglion cells
(parasol, midget, and bistratified) projects to a particular
Internuclear ophthalmoplegia: a conjugate palsy resulting
from damage to the brain stem region known as the medial
longitudinal fasciculus

Sensation and Perception118
layer of the LGN. The parasol retinal ganglion cells pro-
ject to the magnocellular layer of the LGN (forming the M
pathway), the midget retinal ganglion cells project to the
parvocellular layer of the LGN (forming the P pathway), and
the bistratified retinal ganglion cells project to the konio-
cellular layer of the LGN (forming the K pathway). The supe-
rior colliculus can be found at the top of the brain stem,
just beneath the thalamus. Its main function is the control
of rapid eye movements.
4.3
Explain the nature of the retinotopic organization
of V1 and the organization of V2.
After leaving the LGN, the next synapse in the visual path-
way is in the primary visual cortex in the occipital lobe of the
brain. The primary visual cortex is also known as V1 (and area
17). The cerebral cortex is the outer surface of the brain. It
consists of four lobes: the frontal, the temporal, the parietal,
and the occipital. The occipital lobe is the visual brain. V1 is
in the occipital lobe right toward the back of the head. It is
a six-layered structure with different layers receiving input
from different regions of the LGN. V1 is organized in retino-
topic coordinates, which also refer to the spatial organiza-
tion of the external world. One of the features in V1 is cortical
magnification. Cortical magnification means that there is
more space in the cortex devoted to some sensory receptors
than to others. In this case, the fovea has a larger cortical
area than the periphery.
The Nobel prize–winning scientists Hubel and Wiesel dis-
covered both simple cells and complex cells in V1, which are
sensitive to different properties of the visual stimulus. Simple
cells are V1 neurons that respond to stimuli with particular
orientations to objects within their receptive field. The pre-
ferred orientation of a simple cell is the stimulus orientation
that produces the strongest response from the simple cell.
Complex cells are also neurons in V1 that respond optimally
to a stimulus with a particular orientation. But, unlike sim-
ple cells, they respond to a variety of stimuli across differ-
ent locations. In particular, they also respond best to moving
stimuli. End-stopped neurons respond to stimuli that end
within the cell’s receptive field. V1 is organized into columns,
which are sensitive to orientation and ocular dominance. A
hypercolumn is a 1-mm block of V1 containing both the ocular
dominance and orientation columns for a particular region in
visual space. Hypercolumns also include information about
where in space those columns are. Blobs are groups of neu-
rons within V1 that are sensitive to color, whereas interblobs
are the groups of neurons that are sensitive to orientation.
The M pathway, K pathway, and P pathway all enter the cor-
tex through V1. There are three distinct regions within V2,
which match directly up with the three different types of
cells in V1. Thin stripes have color responses, thick stripes
are sensitive to motion, and interstripes are sensitive to
shape and position.
4.4
Compare the difference between the dorsal
pathway and the ventral pathway.
Starting with V1, we can see two distinct pathways, the dor-
sal and ventral pathways. The dorsal pathway is responsible
for information about movement and continues through MT
in the occipital lobe and then into the parietal lobe. The ven-
tral pathway is responsible for object and color recognition.
It goes through areas such as V2 and V4 before heading into
the inferotemporal cortex in the temporal lobe. Nowhere
in the brain has it been found where all of the visual infor-
mation converges to a single location. It now appears that
vision happens with the simultaneous or approximately
simultaneous activation across all of the visual areas.
Forward and backward connections are thought to play a
role in synchronizing a response in the brain, so that our
perceptual experiences are of whole objects and not frag-
mented parts.
4.5
Interpret the concepts of blindsight and conju-
gate gaze palsy.
In the Exploration section, we discussed blindsight, the neu-
ropsychological condition in which damage to V1 causes
a blind spot or scotoma. However, because of noncortical
areas such as the superior colliculus, the patient can still
make visual responses in the absence of conscious seeing.
Blindsight has drawn significant interest from philosophers.
For them, the intriguing phenomenon is that someone is blind
but is making responses to visual stimuli. This seeming para-
dox is resolved when we think of conscious seeing as being
a function of V1 in the cortex.
In the Application section, we discussed conjugate gaze pal-
sies, a series of disorders of vision, brought about by damage
to the oldest part of the brain, namely, the brain stem. In conju-
gate gaze palsy, brain damage to the brain stem occurs that
affects the ability of our eyes to train on the same object.
Conjugate palsies are sometimes seen as complications
in Parkinson’s disease, after cardiac surgery, and when
the pineal gland is compromised. Parinaud’s syndrome
occurs from lesions in the superior colliculus. Rather than
irregular movements and difficulties coordinating the two
eyes, patients with Parinaud’s syndrome have a paralysis of
gaze. They must move their heads, rather than their eyes, to
track motion.

Chapter 4: Visual System: The Brain 119
REVIEW QUESTIONS
1. What is the optic chiasm? Describe how retinal gan-
glion cells cross over in the optic chiasm. How does
this crossing over affect the representation of the
visual world in the LGN?
2. What are the three types of layers in the LGN?
What kind of retinal ganglion cell innervates
each? What is the functional significance of each
layer?
3. What is the function of the superior colliculus?
Where in the brain is it found?
4. What is meant by the term cortical magnification?
How does it explain the retinotopic organization of
V1 of the visual cortex?
5. Describe the receptive fields of simple and complex
cells in V1. What is the function of each type of cell?
6. What are ocular dominance columns? How were
they discovered by Hubel and Wiesel? What role do
they play in the hypercolumns of V1?
7. What are V2 and V3? Where are they found in the
brain, and what is their functional role in vision?
8. What is the dorsal pathway? What is the ventral path-
way? What is the functional significance of each
pathway?
9. What is blindsight? Describe its expression in neuro-
psychological patients. What is the likely anatomical
cause of blindsight?
10. What is conjugate gaze palsy? What is the neurolog-
ical damage in this condition?
PONDER FURTHER
1. Consider the dorsal and ventral pathways and their
perceptual specialties. What might a hypotheti-
cal brain region look like that combined information
from the two systems? What kinds of cells would you
expect to find? What kinds of connections to other
areas of the brain might you see?
2. Scientists have studied receptive fields in very technical
ways, by passing bars of light in front of the eyes of perceiv-
ers. How might these receptive fields respond when pre-
sented with the varied stimuli that occur in the real world?
Would they still respond to the same features, or might
receptive fields actually be tuned to more complex stimuli?
KEY TERMS
Bistratified retinal ganglion cells
(K cells), 95
Blindsight, 114
Blobs, 105
Complex cells, 103
Conjugate gaze palsies, 116
Contralateral organization, 93
Contralateral representation of
visual space, 93
Cortical magnification, 101
Dorsal pathway, 107
End-stopped neurons, 104
Extrastriate cortex (secondary
visual cortex), 106
Hypercolumn, 105
Inferotemporal cortex, 109
Interblobs, 105
Internuclear ophthalmoplegia, 117
Ipsilateral organization, 93
Koniocellular layers, 94
Koniocellular pathway (K pathway), 96
Lateral geniculate nucleus, 94
Magnocellular layers, 94
Magnocellular pathway
(M pathway), 96
Midget retinal ganglion cells
(P cells), 95
MT (V5), 110
Object agnosia, 109
Ocular dominance column, 104
Optic chiasm, 92
Optic tract, 93
Orientation column, 104
Orienting tuning curve, 103
Parasol retinal ganglion cells
(M cells), 95
Parvocellular layers, 94
Parvocellular pathway (P pathway), 96
Primary visual cortex (V1), 100
Retinotopic map, 101
Saccades, 99

Sensation and Perception120
Scotoma, 114
Simple cells, 103
Smooth-pursuit eye movements, 98
Superior colliculus, 98
Ventral pathway, 107
V2, 106
Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e
SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review,
study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning.
Learning Objectives Digital Resources
4.1 Identify the anatomy of the optic chiasm and how that affects
the lateralization of vision in the brain.
Visual Pathways
4.2 Diagram the anatomy of the lateral geniculate nucleus and
the superior colliculus, and describe their roles in visual
processing.
Visual Consciousness Revisited: Magnocellular
and Parvocellular Contributions to Conscious and
Nonconscious Vision
4.3 Explain the nature of the retinotopic organization of V1 and the
organization of V2.
Ocular Dominance Columns: Enigmas and Challenges
Confuse Your Illusion: Feedback to Early Visual Cortex
Contributes to Perceptual Completion
Hubel and Wiesel
4.4 Compare the difference between the dorsal pathway and the
ventral pathway.
Functional Streams and Cortical Integration in the Human
Brain
The Thatcher Illusion Reveals Orientation Dependence in
Brain Regions Involved in Processing Facial Expressions
4.5 Interpret the concepts of blindsight and conjugate gaze palsy. Some Veterans Return With “Hidden” Vision Problems
Blind Man Sees With Subconscious Eye
Pawan Sinha: On How Brains Learn to See

Richard Kail/Science Source

5Object Perception
SSPL/Hulton Archive/Getty Images
LEARNING OBJECTIVES
5.1
Interpret the computational difficulties the visual system must overcome in
recognizing objects as themselves under a multitude of situations and angles.
5.2
Assess the difference between top-down processing and
bottom-up processing and how they affect object perception.
5.3 Describe the gestalt laws of perceptual grouping.
5.4 Explain why the ventral pathway is critical for object perception.
5.5
Discuss how we can distinguish human faces from doll
and mannequin faces, and interpret how object perception
research can inform the airport screening process.
INTRODUCTION
Dr. P. was a professor of music at a college in New York City (Sacks, 1985). His wife
and his colleagues were concerned that he was having problems seeing, though these
problems were not the ordinary problems, such as presbyopia that eyeglasses can cor-
rect. Dr. P. failed to recognize both longtime colleagues and new students. Only when
his acquaintances spoke did he know who they were. In the street, he might stop to
talk to a fire hydrant and be surprised when it did not reply. Finally, his wife had him
visit Oliver Sacks, the noted neurologist and writer, and Sacks wrote of this case in his
famous book The Man Who Mistook His Wife for a Hat (Sacks, 1985). Sacks deter-
mined that Dr. P. had lost none of his intelligence or any of his musical ability. However,
he had a form of agnosia we now know as object agnosia. Agnosias are acquired sen-
sory deficits that occur without any loss of sensation. Dr. P. could not identify simple
objects, despite the fact that he could see them. For example, Dr. Sacks handed Dr. P. a
rose, and Dr. P. responded by telling him that it was “about six inches in length, a con-
voluted red form with a linear green attachment” (p. 13). When given a glove to inspect,
Dr. P. responded by describing it as “a continuous surface, infolded on itself. It appears
to have five outpouchings . . . It could be a change purse” (p. 14). In each case, Dr. P.
demonstrated that he saw the object and could describe it accurately, but he had lost
the ability to identify it, despite what would be to us the obvious nature of each object.
Whereas others might not be able to describe these objects in such eloquent terms,
most people would instantly see a rose and a glove. Although Sacks’s interests lie in the
ISLE EXERCISES
5.1 Segregation and
Grouping
5.2 Ambiguous Figure–
Ground Perception
5.3 Figure–Ground
Symmetry
5.4 Gestalt Laws
5.5 Necker Cube
5.6 Illusory Contours
5.7 Geons
5.8 Facial Responses
in the Brain
(Inferotemporal Region)

124 Sensation and Perception
philosophical interpretations of such a condition, we now know that such descriptions
mark a case of object agnosia, which is likely the result of damage to the inferotemporal
cortex (see Figure 4.19).
What object agnosia shows is that we have specialized areas in the brain that
are necessary for identifying specific objects. Dr. P. sees the shape of the flower, but
he cannot identify it as such. His color perception is accurate, his shape perception
is accurate, but he cannot put them together to see the rose, something any normal
sighted person does effortlessly (Figure 5.1). This suggests that perceptual processes
we take for granted are actually quite complicated. Although it may be difficult to
imagine what it is like to be agnosic, we hope to show you how complex the pro-
cesses are that allow us to effortlessly visualize objects. Examine the two photo-
graphs in Figure 5.2. Both depict the same TV remote, each from a different angle.
The representation of the remote is very different on the retina in each photograph.
In Figure 5.2a, we are looking directly down on the remote. It is flat against the sur-
face, and we can see only one side. In Figure 5.2b, the remote is on its side, and we
are viewing straight on rather than from above. Yet most of us will have no difficulty
seeing the same object. This is an example of shape constancy, the concept that an
object remains the same despite changes to its retinal image. How the visual system
recognizes objects despite such changes is one of the topics we explore in this chapter.
INTRODUCTION TO OBJECT PERCEPTION
5.1
Interpret the computational difficulties the visual system must overcome in
recognizing objects as themselves under a multitude of situations and angles.
Despite Dr. P.’s intelligence and musical ability, he probably would not be able to func-
tion in everyday life if he did not have his wife to guide him through it. Think of trying
to prepare a meal for oneself when one mistakes one’s newspaper for the frying pan,
and one’s telephone for the spatula. It would be a real disaster and quite dangerous.
Indeed, under such circumstances, it might be wiser for Dr. P. to blindfold himself and
rely exclusively on his other senses, rather than trust his distorted vision. Dr. Sacks
never followed up on his patient, so we never learned what happened to him after their
initial meeting. Nonetheless, Dr. P.’s dilemma illustrates an important concept for the
rest of us. Being able to recognize quickly the
objects in our environment is critical.
Consider the environment you are in now.
Surrounding you are many objects. Most
likely you have a book or laptop (or tablet
device) in front of you. If you have come this
far in the course, this textbook has already
become a familiar if hefty companion. If you
are reading it on your tablet computer, that
object was already familiar to you before you
started the course. Perhaps you are drinking
coffee from your favorite mug. Your eye-
glasses and phone rest on the table near you.
You are sitting in the comfortable chair that
your aunt purchased for you last year. You
are surrounded by the familiar objects that
you recognize instantly. However, our ability
FIGURE 5.1 Roses
Convoluted red forms or roses?
What do you see?
©
iS
to
ck
ph
ot
o.
co
m
/D
N
Y5
9
FIGURE 5.2 Two Perspectives of the Same Object
The same object looks very different from different angles and creates very
different images on the retina. But those of us with normal visual perception
have no difficulty deciding that these photographs depict the same object.
(a) (b)

125 Chapter 5: Object Perception
to recognize objects goes far beyond that with which we are immediately familiar. A
friend comes in with a new mug, and we recognize the mug instantly as a mug, and we
do not have to ask what such an object is. Another friend enters with the latest cellular
phone, and we recognize it as a cell phone even if we have never seen that particular
model before. In fact, we recognize individual unfamiliar examples of familiar catego-
ries all the time—a new species of tree, a paper clip shaped like a musical note. Only
occasionally do we come across a genuinely unfamiliar object—one about which we
have no knowledge of what it is made of, what its function is, or where it came from
(Figure 5.3). We see here how much knowledge influences our perception. When we see
a genuinely unfamiliar object like this, even its size may not be clear.
However, even recognizing a familiar object is an enormously difficult compu-
tational problem (however effortless it feels to us). Consider Figure 5.4. In order to
recognize the television remote in this picture (and yes, it is the same as in Figure 5.2),
our visual systems must do a great deal of computation. Consider first that the top
half of the remote is being illuminated by the light from a nearby window and, as a
consequence, is reflecting significantly more light than the bottom half, which is in
the shadow of the wall. Nonetheless, we see the remote as a continuous object. The
pen in the glass bowl is obscuring part of the remote. Finally, the glass bowl is dis-
torting the image of the far upper corner of the remote. But what we see is a standard
TV remote, not some odd space-age object. The point here is that our visual system
has to make a lot of complex inferences for us to see even an ordinary object in a
familiar scene.
Our ability to detect objects must overcome three aspects of the environment:
image clutter, object variety, and variable views. To overcome image clutter, we
must discern the object despite the overlapping presence of nearby objects, such
as coasters, other remotes, and glass bowls. Although detection of objects is nor-
mally easy, increased clutter can make it difficult, as in the Where’s Waldo? series
of children’s books. With respect to object variety, we must recognize a particu-
lar object as a member of a particular class or
category despite small or large differences from
our general prototype of that category. That is,
the remote in Figure 5.4 is likely one that you
have never seen before. However, because of its
similarity to other remotes, chances are you had
no difficulty assigning it to the correct category.
Finally, in variable views, we must recognize an
object despite its being placed in very different
orientations relative to ourselves and casting
different images onto our retinae. That is, we
must be able to identify the same object from
different vantage points (as in Figure 5.2). This
is particularly important when it comes to face
recognition. You want to be able to recognize
your friend regardless of the angle at which you
are looking at him or her. The point here is that
object recognition actually is a computationally
difficult task for our visual systems.
Another way of examining the idea that
object recognition is a hard problem is to look
at the development of self-driving cars (see Cain
Miller & Wald, 2013; Findling, 2017). Although
self-driving cars have come a long way, no state
FIGURE 5.3
A Genuinely Unfamiliar Object
Can you tell what this object is?
Can you judge its approximate
size? What function might this
object have? What is it made
of? When you do not have
knowledge about the object,
these questions may be very
tricky to answer.
©
Shutterstock.com
/Charles Curtis
FIGURE 5.4 A Complex Scene
Despite the presence of odd lighting, obstructions, and a clutter of
other objects, we can still recognize the remote behind the glass bowl.
Our visual systems are tuned to identify objects across a wide range of
transformations.

126 Sensation and Perception
allows them unless there is also a human driver
sitting at the wheel. Why do we still not trust
self-driving cars? One reason is the difficulty of
object recognition. The computer that responds
to the video input must be able to quickly and
accurately recognize many thousands of objects
from any number of angles. More important than
distinguishing between two views of a remote,
the self-driving car must distinguish between dis-
carded dolls, which may not require swerving, and
children on the street, which must be avoided at all
costs. At present, according to Findling (2017), the
technology is finally there, though issues of legal
responsibility for accidents remain wide open.
Self-driving cars, of course, have certain advan-
tages over human drivers: They do not get tired,
they do not get angry or feel road rage, they do not
text while driving, they do not drink alcohol, and
they can be programmed not to tailgate. Thus, as
object recognition software improves, it is likely
that self-driving cars will ultimately become much
safer than human driving (Figure 5.5).
TEST YOUR KNOWLEDGE
1. What is object agnosia? How is object agnosia different than blindsight?
2. What computational difficulties does the visual system face in recognizing objects? How
are these difficulties related to the difficulty in building self-driving cars?
TOP-DOWN PROCESSING AND
BOTTOM-UP PROCESSING
5.2
Assess the difference between top-down processing and bottom-
up processing and how they affect object perception.
In sensation and perception research, we seek explanations both at the neurological level
and at the psychological level. We need to understand what is going on in the sensory
organs and with the brain in order to make sense of sensory processes, but we also need
to understand psychological processes in order to determine how we go from sensation
of a pattern of light to the recognition of objects, such as individual people, roses, and
self-driving cars. To this point, we have mostly been discussing the physiology and anat-
omy of sensation, but in this chapter, we begin discussing the more psychological issues
of perception. One way to view this distinction is in the cognitive psychology logic of
top-down and bottom-up processing. Bottom-up processing means that physical stim-
uli influence how we perceive them. This is relatively straightforward: Our eyes detect
reflected long-wavelength light when we examine the object depicted in Figure 5.1, and,
therefore, we see it as red. Top-down processing means that our existing knowledge of
objects influences how we perceive them. Thus, we perceive a “rose” because we already
know that roses are bright red and have this particular pattern of shape and shadow. If we
did not know about roses, we might give a description similar to the one given by Dr. P.
©
B
ro
ok
s
Kr
af
t/
Co
rb
is
N
ew
s/
G
et
ty
FIGURE 5.5 A Self-Driving Car
This is a photo of Google’s prototype of a self-driving car. The computer
that navigates the car must make real-time computations on a dizzying
amount of data to negotiate the road and avoid obstacles. These
computations are done routinely in the human brain but are difficult to
program into a computer.
Bottom-up processing: a
process whereby physical
stimuli influence how we
perceive them
Top-down processing: a
process whereby our existing
knowledge of objects influences
how we perceive them

127 Chapter 5: Object Perception
Look back again at the object depicted in Figure 5.3. Because we do not know
what the object in Figure 5.3 is, we have a hard time perceiving many of its important
aspects, such as its size. Could you fit this object in your hand, or could you fit yourself
inside the object? It is hard to tell. As you read through this chapter, think about when
the issue at hand is bottom-up processing and when it is top-down processing.
Recognition and Representation
Although the focus of this chapter is object perception, the issues discussed here dovetail
with issues of memory, in this case our knowledge base of visual objects. As we just dis-
cussed, our existing knowledge of the world is important in identifying individual objects
in it. Because existing knowledge is important to perceiving objects, it is necessary to have
a memory system capable of storing this knowledge. Thus, memory and perception are
integral to understanding object perception. As
a result, we need to define two memory terms
here. Recognition refers to the ability to match
a currently viewed item with an item in memory.
In more technical terms, it refers to the percep-
tual matching of something currently present to
our visual system with a stored representation
in memory. We can think of recognition as being
equivalent to a police lineup (Figure 5.6). A wit-
ness is shown several possible suspects and must
pick out the one who was seen at the crime. In
perceptual terms, recognition implies that we
know something about what we are seeing—that
is, we are able to match the object we see with a
stored memory. Obviously, recognizing objects is
an important part of our visual perception. In the
case of Dr. P., he sees objects but does not recog-
nize them as members of learned categories, such
as gloves or flowers. In normal vision, we must
recognize two classes of objects. We must recog-
nize specific objects as members of larger classes.
For example, we must recognize the object on the
shelf as a screwdriver. But we also must recognize objects as specific instances of that cat-
egory. For example, that screwdriver on the shelf is my favorite screwdriver. Being able to
recognize specific instances of a particular category is vital in face recognition, in which you
must recognize a person as, for example, your uncle, rather than as a member of a category,
such as men (Schwartz, 2017).
Representation refers to the processes that translate stimulus information into a per-
ceptual experience of that stimulus. Thus, an object appears in the visual field. Neural
processes detect it and send an image to V1. The process of representing that object
starts immediately, in parallel across the dorsal and ventral system. The representation
then is the neural code for what the object actually is. This use of the term is not to be
confused with other meanings of the term representation in related fields. In neurosci-
ence, representation refers to how neurons code for information, such as the shapes
and colors of objects. It can also denote how networks of neurons across the brain
store the information we will need to retrieve at some point. In cognitive psychology, a
representation is the form in which information itself is stored. If you consider a class of
objects, such as familiar faces, we need to be able to represent information about them.
We need to be able to represent the general form of human faces and the specific forms
of particular human faces (e.g., your mother, your brother, your landlord).
©
iStockphoto.com
/RichLegg
FIGURE 5.6 Recognition and Representation
Recognition: the ability to
match a presented item with an
item in memory
Representation: the storage
and/or reconstruction of
information in memory when
that information is not in use

128 Sensation and Perception
TEST YOUR KNOWLEDGE
1. What is the difference between top-down processing and bottom-up processing?
2. What processes are necessary in order to be able to remember an object that you
saw earlier?
Perceptual Organization
Perceptual organization is the process by which multi-
ple objects in the environment are grouped, allowing us
to identify those objects in complex scenes. Look around
you. You are likely to see a complex scene. Books, com-
puter equipment, pens, and pencils, as well as other
personal possessions, probably surround you if you are
reading this book at home. Other people, bookshelves,
paintings, and other objects are likely to surround you
if you are reading this at the library. Perceptual organi-
zation is necessary, as almost any scene will be complex,
with overlapping objects, occluding objects, and some-
times ambiguous objects. Perceptual organization per-
mits us to group what we see into coherent perceptions.
That is, it is necessary to group some objects together
(e.g., books) and separate other objects even if they have
similar shapes (books from bricks, e.g.).
Two important processes in perceptual organiza-
tion are grouping and segregation. Grouping is the pro-
cess by which elements in a figure are brought together
into a common unit or object. In Figure 5.4, we see the
remote as one object or grouped together, even though
the glass bowl occludes part of the remote. Segregation
is the process of distinguishing two objects as being
distinct or discrete. Thus, in Figure 5.4, we need to seg-
regate where behind the glass bowl the end of the first remote is located and where
the beginning of the second remote starts. In the formal table scene depicted in Figure
5.7, a perceiver will group the plates together, the chairs together, and the squares on
the tablecloth together. However, segregation will be used to distinguish those plates
that will contain people’s individual meals and the serving dishes that will be spared.
You can see an illustration of these principles on ISLE 5.1.
GESTALT PSYCHOLOGY AND
PERCEPTUAL ORGANIZATION
5.3 Describe the gestalt laws of perceptual grouping.
In the early 20th century, while behaviorists were ascendant in the United States, a school
of psychology known as gestalt psychology pursued a different path in research, largely
in Western Europe, particularly Germany. The gestalt psychologists were led by such
psychological greats as Max Wertheimer, Kurt Koffka, and Wolfgang Köhler. In a general
sense, the gestalt theorists claimed that the brain is holistic, with self-organizing tenden-
cies. The higher levels of organization take precedence over the lower levels. This differs
Perceptual organization: the
process by which multiple objects
in the environment are grouped,
allowing us to identify multiple
objects in complex scenes
Grouping: the process by
which elements in a figure are
brought together into a common
unit or object
Segregation: the process of
distinguishing two objects as
being distinct or discrete
Figure–ground organization:
the experience viewers have as to
which part of an image is in front
and which part of an image is in the
background of a particular scene
ISLE 5.1
Segregation and Grouping
FIGURE 5.7 Fancy Dinner
How many distinct candles are there on the table? You probably
have no difficulty determining the number of candles, despite the
visual obstruction of some of the candles. Moreover, it is trivial
for you to determine where one object ends, such as the chairs in
the front, and where the table begins. Although easy in practice,
this is a complex process computationally and relies on stored
representations of objects.
©
iS
to
ck
ph
ot
o.
co
m
/a
lo
ha
_1
7

129 Chapter 5: Object Perception
to some extent from current research trends, which are often
reductionist, in the sense that we try to build our models from
the lowest levels upward. In terms of vision, gestalt psychol-
ogy argues that what we see is greater than its individual parts.
That is, the processes of perception are designed to see the scene
rather than bits of light here and there. In this way, gestalt psy-
chologists thought that stimuli were sufficiently rich in structure
to allow the perceptual system to extract meaning directly from
the stimuli, rather than building it up from image to thought.
As such, it was important for them to establish the rules that
governed the building of bigger perceptual units.
The gestalt psychologists were structuralists who thought
that conscious perception rested on the building blocks of
sensation. However, unlike Helmholtz, gestalt psychologists
thought that stimuli were rich enough for meaning to be inter-
preted directly rather than through unconscious inference.
Thus, we sense incoming stimuli, but perception is an active
process that interprets perceptions out of these sensory build-
ing blocks. Gestalt psychology also maintained that the whole
is often different from the sum of its parts. That is, the percep-
tion that emerges from a physical scene may not be directly
predicted by the sensory components that it is composed of;
rather, it emerges when we integrate the components into a
whole. Think of the standard way in which comic books
used to be created. The images were composed of many col-
ored dots. Each dot was just a little speck of red, but when put
together, one saw Spiderman. Examine Figure 5.8. The eye in
Figure 5.8 is composed of many dots. To gestalt psychologists,
the dots make up the sensory components of the stimuli, but
seeing the eye emerges from perception. Moreover, the stimulus
makes no sense unless we see it as an eye. It is just dots other-
wise. Thus, this immediate perception of higher level interpreta-
tions is necessary in any visual system.
For these reasons, gestalt psychologists were interested in
understanding the rules by which perception picked out the
whole from its parts. In particular, they investigated three prin-
ciples concerning how our perceptual systems do this. These
principles are figure–ground relations; the laws of Prägnanz, or
good fit; and the laws of grouping. Although these ideas were
developed approximately 100 years ago, they still apply to the
science of vision in important ways. We start with figure–ground
organization.
Figure–Ground Organization
Figure–ground organization refers to the experience viewers have
as to which part of an image is in the foreground and which part
is in the background of a particular scene. In essence, we divide
the world into two elements: the figure that is the object of regard
and the rest, which is ground or background (here they mean the
same thing). In many cases, we can choose what will be the figure.
For example, if you are giving a public lecture, you can look at
©
iStockphoto.com
/jc_design e
FIGURE 5.8 Eye or Dots?
All that is really present physically in this image is a
series of dots on a background. However, the arrangement
of the dots creates the image of an eye. Gestalt
psychologists argue that the process behind seeing the
eye instead of the dots is fundamental to perception. Our
visual systems extract this higher meaning from the array
of stimuli in the world.
FIGURE 5.9 Figure–Ground
If you are looking at the cat, other aspects of the stimulus
become ground or background. However, if you switch
your focus from the cat to the oranges, the cat becomes
background.

130 Sensation and Perception
one person’s face in the audience. As you look at that
face, that face is the figure, and everyone else at that
moment is ground. When you look at a new face, the
last person’s face is now ground. Consider Figure 5.9.
When you look at this photograph, when you direct
your gaze toward the oranges, the cat is ground and
the oranges are figures. However, when you look at the
cat, figure–ground reverses. In other situations, figure
and ground may be determined by other components
of the stimulus. If you look around, you will see objects
close to you and objects farther away. Objects close
to you tend to be lower in your visual field and less
obstructed by other objects than those farther away. Is
such information useful to your visual system? Indeed
it is, as normally, figure–ground determinations are rel-
atively easy, as is apparent in Figure 5.10. In this figure,
we tend to see the puffins as the foreground and the sea
as the background.
However, there are exceptions in which the bot-
tom of an image may not represent what is close, and
the top of an image may not represent what is far away. Consider the photograph in
Figure 5.11. Here we are looking down and across at a pond in the Everglades. Here,
the figure circles the background. There are branches hanging down from the top of
the image that are close to the observer in addition to the vegetation at the bottom,
which is also in the foreground. The background of the photograph is the stuff in the
middle, which is farthest from the observer. Because of this, this photograph takes
a moment or two to figure out, because it does not conform to ordinary rules of
figure–ground. Note the turtle swimming just beneath the surface of the pond at the
center of the photograph.
©
R
oi
ne
M
ag
nu
ss
on
/P
ho
to
di
sc
/G
et
ty
Im
ag
es
FIGURE 5.10 Figure–Ground
In this figure, we tend to see the resting puffins as the figure, because
they are closer to us, and the distant sea as background.
FIGURE 5.11 Figure–Ground
In this photograph of an Everglades alligator hole, the construction of the
scene leads one to see the middle as foreground and the top, bottom, and
sides as background.
FIGURE 5.12 More on Figure–Ground
Do you see two faces looking at each other in front of an
orange background? Or do you see an orange vase in front of
a white background? You may see one or the other and switch
between the two, but you cannot see both at the same time.

131 Chapter 5: Object Perception
There are other situations in which figure–ground can be highly ambiguous, and
it may not be possible to determine which part of the image is the front and which is
the back. This is true of many classic visual illusions. One such classic example is the
face–vase figure, first introduced into the literature by psychologist Edgar Rubin in
1915 (Figure 5.12). In this figure, the border between the orange and white regions
is seen as being part of one or the other, not both. Thus, either the orange vase at the
middle stands out as the foreground and the white faces are in the background, or
the white faces stand out as the foreground and the orange space is the background.
Interestingly, in these ambiguous figures, people see either one interpretation or the
other, and although they can flip back and forth, they cannot see both at the same
time. You can see a demonstration of this illusion on ISLE 5.2.
One of the useful aspects of such illusions is that we can make systematic changes
in the way the images are presented and see how those changes affect the perception
of the figure. Consider Figure 5.13. In this figure, we see that the field representing
the faces and the vase are offset. When the vase is lowered and moved to a corner, it
stands out as the figure, and it is nearly impossible to see the faces. When the faces
are lowered and moved to the corner, they stand out as the figure, and the vase is no
longer visible in the background. This illustrates the point that figures tend to be in
the front and the background on top.
A Few Rules That Govern What We See as
Figure and What We See as Ground
We have already seen that the top of an image is more
likely to be seen as the background and that what is in
the front of an image is likely to be seen as the figure.
But there are other rules as well that apply to the figure–
ground relation. We outline them here (also see ISLE 5.3).
We can identify another environmental feature that
influences our perception of figure–ground. As we have
seen already, perceived depth is one feature. The figure
that appears in the foreground is often below the fig-
ure that appears in the background. Another feature is
symmetry. A figure with symmetrical borders is more
likely to be judged as being in the foreground than in
the background (Figure 5.14).
ISLE 5.2
Ambiguous Figure–Ground
Perception
FIGURE 5.13 The
Face–Vase Figure–Ground
Illusion
(a) With the shifting down of the
position of the orange area, we
are more likely to see the vase.
(b) With the shifting down of the
position of the white area, we
are more likely to see the faces.
(a)
ISLE 5.3
Figure–Ground Symmetry
FIGURE 5.14
Symmetry Affects Figure–Ground Organization
Images that are symmetrical are more likely to be seen as figure
and therefore in the foreground, whereas less symmetrical images
are more likely to be perceived as background.
(b)

132 Sensation and Perception
Another feature that affects figure–ground percep-
tion is that a figure is more likely to be perceived as
being in the foreground if it is perceived to be on the
convex side of a border. That is, the figure that appears
to be in the foreground is the one with outward bulg-
ing (convex) borders, not the one with inward-facing
(concave) borders. This point is illustrated nicely in
an ambiguous figure taken from research by Stevens
and Brookes (1988). In Figure 5.15, you can see either
swirly ropes descending or a thorny branch descend-
ing. Most perceivers, though, see the swirly ropes, as
these are the convex figures.
That convex images are perceived as figure and
concave images as ground was also illustrated in an
experiment by Peterson and Salvagio (2008). These
researchers presented figures like those depicted in
Figure 5.16. In Figure 5.16, one can see convex blue bands and concave white bands.
According to the ideas concerning convexity and figure–ground relations, we should
see the blue bands as the figure and the white bands as the background. Peterson and
Salvagio placed a small red square at some point on the figure and had participants
make judgments of front or back. When the square landed on a convex band, partic-
ipants indicated that it was in the front, but when the square was on a concave band,
they indicated that it was in the background.
In summary, there are a number of rules that govern figure–ground relations. These
rules allow us to make sense of complex scenes because we use our existing knowledge
of the world to help us interpret what is visually in front of us. For example, we know
from experience that heavy objects do not usually float 3 feet above the ground. Thus,
even though we cannot see the bottom of a chest of drawers, we infer that there must
be a bottom, and that the couch we see in its entirety is in front of the chest of drawers.
Thus, the couch becomes the figure and the chest is ground behind it. We next turn to
more of these rules developed by the gestalt psychologists.
Gestalt Laws of Perceptual Grouping
The process by which visual systems combine figures within an image into wholes is called
perceptual grouping. Perceptual grouping involves using existing knowledge to place simi-
lar items together or to group images in different parts of the visual field into a perception
of the same object. Consider Figure 5.4 once more. When you view this image, you imme-
diately group figures into coherent wholes. Thus, even though the blue on the two coasters
is the same, you see these as two differ-
ent objects because they are not contin-
uous; one blue pattern is associated with
the “pig” coaster, whereas the other blue
pattern is associated with the “rooster”
coaster. This is also why we see a complete
glass bowl even though it is possible that
the bowl is broken, as the reflection of sun-
light makes it difficult to see the complete
bowl.
The gestalt psychologists, especially
Max Wertheimer, developed a number
of “laws” that predict how perceptual
FIGURE 5.15
Convexity Affects Figure–Ground Organization
Images with convex borders are more likely to be seen as figure,
whereas those with concave borders are more likely to be seen as
ground. Adapted from Stevens and Brookes (1988).
FIGURE 5.16 Figure–Ground Judgments
The participants’ task was to decide if the red square was on the figure or in the
background. When the square landed on a convex band, participants indicated
that it was in the foreground, but when the square was on a concave band, they
indicated that it was in the background. Based on stimuli from Peterson and
Salvagio’s (2008) experiment.

133 Chapter 5: Object Perception
grouping occurs under a variety of circumstances (Wertheimer,
1923/1938). Technically, in sciences, “laws” are predictions
that are always true, no matter the conditions. In reality, these
laws are better classified as principles, which are true most of
the time. However, the term laws has stuck with these prin-
ciples because they were established so long ago. So we will
continue to refer to them as laws, even though technically they
are not scientific laws.
The law of good continuation means that figures that have
edges that are smooth are more likely to be seen as continuous
than those with edges that have abrupt or sharp angles. This
law is best seen rather than described. Examining Figure 5.17a,
most people see a horizontal orange bar behind a blue vertical
bar. This is because the two orange rectangles would meet if
each was extended. So we see them as continuous behind the
blue vertical bar. However, in Figure 5.17b, the two orange
rectangles would not be continuous behind the blue vertical bar. Therefore, we see
them as two separate entities.
The law of proximity is the gestalt grouping law that states that elements that are
close together tend to be perceived as a unified group. This straightforward law states
that items close to one another tend to be grouped together, whereas items farther
apart are less likely to be grouped together. In Figure 5.18, we tend to see five groups
of 3 letters rather than one group of 15 letters.
The law of similarity is the gestalt grouping law that states that elements that are
similar to one another tend to be perceived as a unified group. Similarity can refer to
any number of features, including color, orientation, size, or, indeed, motion. This is
illustrated in Figure 5.19.
The law of symmetry is the gestalt grouping law that states that elements that are
symmetrical to one another tend to be perceived as a unified group. Similar to the law
of similarity, this rule suggests that objects that are symmetrical to one another will
FIGURE 5.17 Good
Continuation
(a) We see one bar on top
of another bar. The good
continuation of the bar in the
back suggests its presence.
(b) We see two separate bars
behind the front bar because
the break in the figure does not
suggest continuity.
(a)
(b)
FIGURE 5.18 Proximity
When we see these letters, we
see five groups of three letters
because of the proximity of the
three letters together. That is,
elements that are close together
tend to be grouped together.
FIGURE 5.19 The Law of Similarity
We tend to group together objects that look similar. Thus, we see objects grouped together by
color (c) and objects grouped together by size (d).
(a) No grouping
(b) Proximity
(c) Similarity of color
(e) Similarity of orientation
(d) Similarity of size
(f) Common fate
Law of good continuation:
the gestalt grouping law stating
that edges that are smooth
are more likely to be seen as
continuous than edges that
have abrupt or sharp angles
Law of proximity: the
gestalt grouping law stating
that elements that are close
together tend to be perceived
as a unified group
Law of similarity: the gestalt
grouping law stating that
elements that are similar to one
another tend to be perceived as
a unified group
Law of symmetry: the
gestalt grouping law that
states that elements that are
symmetrical to each other tend
to be perceived as a unified
group
LKS AJC QWE PKX JPB

134 Sensation and Perception
FIGURE 5.20 The Law of Symmetry
In this figure, the figures that are
symmetrical are grouped together.
Symmetry
be more likely to be grouped together than objects not symmetrical to one another.
This is illustrated in Figure 5.20.
The law of common fate is the gestalt grouping law that states that
elements that are moving together tend to be perceived as a unified
group. Think of watching a flock of geese moving across a fall sky.
The geese are all flying in the same direction at approximately the
same speed. Therefore, we see them as a gestalt group or, in this case,
a flock. We might see a second flock flying in a different direction in
a different pattern. You can see an illustration of each of the gestalt
principles on ISLE 5.4.
Perceptual Interpolation
In real-world scenes, as we have already mentioned, particular objects are often par-
tially occluded. That is, a particular object may be blocked by another object, rendering
the physical image of the first object disjointed. Yet our perceptual systems see objects
as continuous wholes despite that blocking. In Figure 5.21a, for example, the leaves
of the tree do not impede our perception of a single koala, even though parts of the
koala are not visible. In Figure 5.21b, the iron gate breaks up our view of the grass
field beyond, but we see a continuous field of grass rather than a series of smaller ones
divided by some unseen fence that coincides exactly with the grates of the gate. Look
around you. Many of the objects around you are occluded by other objects, yet you see
each object as a whole rather than a set of disjointed parts. Thus, filling in edges and
completing surfaces are important jobs of our object recognition system.
One important process used in perceptual interpolation is edge completion.
Edge completion is the perception of a physically absent but inferred edge, allow-
ing us to complete the perception of a partially hidden object. A stunning exam-
ple of edge completion comes from illusory contours, as illustrated by the famous
ISLE 5.4
Gestalt Laws
Law of common fate: the
gestalt grouping law that states
that elements that are moving
together tend to be perceived
as a unified group
Edge completion: the
perception of a physically
absent but inferred edge,
allowing us to complete the
perception of a partially hidden
object
©
iS
to
ck
ph
ot
o.
co
m
/P
ix
el
-P
ro
du
ct
io
ns
FIGURE 5.21 Perceptual Interpolation
We infer the continuation of objects even when they are partially occluded by other objects. In Figure 5.21a, for example, the leaves
of the tree do not impede our perception of a single koala, even though parts of the koala are not visible. It could be that the koala is
terribly disfigured and the leaves are arranged so that we cannot notice it. But it is more likely that this is a normal super-cute koala. In
Figure 5.21b, the iron gate breaks up our view of the grass field beyond, but we see a continuous field of grass rather than a series of
smaller ones divided by some unseen fence that coincides exactly with the grates of the gate.
©
iS
to
ck
ph
ot
o.
co
m
/V
ik
to
rC
ap (a) (b)

135 Chapter 5: Object Perception
FIGURE 5.22
Kanizsa Triangle
These illusory contours illustrate
the principle of edge completion.
The arrangement of the circles
with the bites taken out of them
suggest that a white triangle
overlaps three blue circles.
Because we unconsciously
infer this, we see the triangle as
slightly brighter in color than the
white background, even though
it is not.
FIGURE 5.23
An Illusory Necker Cube
Perceptual interpretation here
creates the appearance of a
white-barred cube resting on
top of a number of blue circles.
The bars that appear to define
the cube seem brighter than
the surrounding background.
This is due to the nonconscious
inference of the cube. Moreover,
the cube created is ambiguous
and can be seen in one of two
orientations.
ISLE 5.5 and 5.6
Necker Cube
Illusory Contours
Kanizsa triangle (see Kanizsa, 1979;
Petry & Meyer, 1987). Kanizsa devel-
oped these illusions explicitly working
in the tradition of gestalt psychology,
and you can apply many of the princi-
ples of gestalt psychology to interpret
or predict the presence of illusory con-
tours. Illusory contours (or subjective
contours) are perceptual edges that
exist because of edge completion but
are not actually physically present.
In Figure 5.22, we see a bright white
triangle imposed on a background of
blue circles. However, looking closer,
there is no real difference in bright-
ness between the white of the triangle
and the white of the background. In
fact, if you can force yourself to see
three blue Pac-Man figures rather than
three occluded triangles, the illusion
of the bright triangle may disappear.
Important to note here is that most
people really do see a brightness dif-
ference. Only when you block out
the figure and compare the brightness
of the center of the triangle with the
brightness of the background do you
perceive them as being identical. And to really “mess with your mind,” examine the
illusory Necker cube in Figure 5.23 and ISLE 5.5. You can try ISLE 5.6 to control
the parameters yourself and see how they affect the perception of illusory contours.
Illusory contours have interested perception researchers for a long time (see
Petry & Meyer, 1987). This includes neuroscientists as well as gestalt psycholo-
gists. Neuroscience data suggest that illusory contours are processed at a very early
level in vision, consistent with the illusory brightness differences. The physiological
explanation for illusory contours arises from studies on V2 neurons in the brains of
monkeys (von der Heydt, Peterhans, & Baumgartner, 1984). In these studies, von
der Heydt et al. (1984) showed that edge detection cells responded to illusory edges
as strongly as they did to real ones when the stimuli were aligned to create percep-
tual illusions of edges. However, the cells did not fire when stimuli without implied
edges were presented. Other research shows that cells in V1 are activated by illusory
contours in the same way that they are activated by actual contours (Maertens &
Pollmann, 2005). Thus, illusory contours appear to be a relatively low-level feature
of object identification. This reinforces the notion that the gestalt principles are
based on low-level features of perception, consistent with the gestalt view that we
pick up these features in the stimulus rather than as a function of nonconscious
processing.
TEST YOUR KNOWLEDGE
1. What is figure–ground symmetry? Why is it important in visual perception?
2. What are the laws of perceptual grouping? What do they account for?
Illusory contours: perceptual
edges that exist because of
edge completion but are not
actually physically present

136 Sensation and Perception
RECOGNITION BY COMPONENTS
Whereas the gestalt psychologists emphasized organization, that is, how we use implicit
knowledge to help us perceive objects in the environment, other researchers have
emphasized the bottom-up approach, that is, how we use the information in the world
to construct a perception of what we see. One of the most influential bottom-up the-
ories advanced to account for object recognition was developed by Irving Biederman
in the 1980s (see Biederman, 1987). In this view, the complexity of object recognition
is solved when the visual system breaks down objects in the environment into what
Biederman called geometric ions, or geons. Geons represent the basic units of objects
and consist of simple shapes, such as cylinders and pyramids. Recognition by com-
ponents theory states that object recognition occurs by representing each object as a
combination of basic units (geons) that make up that object. We recognize an object
by the relation of its geons. Biederman tentatively proposed that there were roughly 40
independent geons and that just about any object could be represented by some com-
bination of these geons. Figure 5.24 shows a few sample geons, and Figure 5.25 shows
simple objects made up of geons. You can also do an interactive version of building
shapes from geons on ISLE 5.7.
One of the advantages of this model is that if any object is a combination of a
few basic geons, then the object is specified by those components, and its position
relative to the observer should not matter. Thus, this model accounts for the view-
point invariance of objects, that is, that objects are seen as the same regardless of the
vantage point relative to a viewer (see Figure 5.4 again). However, recently recog-
nition by components theory has fallen into disfavor, largely because of its limits in
accounting for some phenomena, such as letter recognition and face recognition, two
big categories of object recognition in humans.
TEST YOUR KNOWLEDGE
1. What are geons, and why are they considered the building blocks of object perception?
FIGURE 5.24 Geons
These basic shapes, or geons, were thought by Biederman
(1987) to be the basic building blocks of object perception.
ISLE 5.7
Geons
Geons: the basic units of
objects, consisting of simple
shapes such as cylinders and
pyramids
Recognition by
components: a theory stating
that object recognition occurs
by representing each object
as a combination of basic
units (geons) that make up that
object; we recognize an object
by the relation of its geons
Viewpoint invariance: the
perception that an object does
not change when an observer
sees the object from a new
vantage point
FIGURE 5.25 Objects From Geons
When we put various geons together, we can create various
recognizable objects.
(a) (b) (c) (d) (e)

137 Chapter 5: Object Perception
THE NEUROANATOMY AND
PHYSIOLOGY OF OBJECT PERCEPTION
5.4 Explain why the ventral pathway is critical for object perception.
We start this section with a brief review of some of the material covered in the last
chapter. Remember that two pathways emerge from area V1 of the occipital cortex.
One of these pathways is known as the “where” pathway or the dorsal pathway. This
pathway works its way through the extrastriate cortex and then continues on to the
parietal lobe. The dorsal pathway is concerned mainly with locating objects in space
and perceiving motion. We focus on this pathway in Chapter 8. The second pathway
that emerges from V1 is known as the “what” pathway or ventral pathway. This path-
way works its way through the extrastriate cortex and then continues on to the tem-
poral lobe. One of its main functions is object recognition. We focus on this pathway
in this section.
Representation of Shapes in Area V4
After information leaves V1 and travels toward the extra-
striate cortex along the ventral pathway, one of the impor-
tant loci is area V4 in the occipital cortex (Figure 5.26). V4
has been linked to color vision but also to shape perception.
V4 neurons seem to have a clear preference for edges, but a
more complex analysis of edges than seen in V1 or V2. For
example, V4 neurons can respond to edges that are either
straight or curved, whereas V1 neurons show a strong pref-
erence for straight edges (Pasupathy & Connor, 2002). Also,
like V1 neurons, V4 neurons respond to contours (Pasupathy
& Connor, 2002). However, V4 neurons respond to either
convex or concave contours. Thus, this area is involved in
delineating shapes, necessary for object recognition.
Object Recognition in the
Inferotemporal Area
Information in the ventral pathway leaves the occipital lobe
and heads into the inferotemporal area of the temporal lobe
(see Figure 5.26). Neurons in the inferotemporal area have
much larger receptive fields than those in V1 and V4 and seem to be devoted to detect-
ing particular kinds of objects anywhere in the visual field rather than specific features
in specific places. Indeed, some areas in the inferotemporal area seem to be very specific
to particular kinds of shapes or forms (Brincat & Connor, 2004). That is, rather than
detecting edges or contours, the inferotemporal area seems to specialize in detecting
specific objects from chairs to bears to faces.
Neuroscience researchers have known for a long time of the relation between the
inferotemporal cortex and object recognition. Back in the 1930s, Klüver and Bucy
(1939) lesioned the temporal lobes of monkeys in their lab. After creating these lesions,
V4: an area of the brain
involved in both color vision and
shape perception
Inferotemporal area: the area
of the temporal lobe involved in
object perception; it receives
input from V4 and other areas in
the occipital lobe
FIGURE 5.26 Object Recognition in the Brain
Area V4 and the inferotemporal area are two critical areas in
the neural substrate of object recognition.
V4
Inferotemporal cortex

138 Sensation and Perception
Klüver and Bucy observed what they called “psychic blindness” but
what we now call object agnosia. The monkeys were able to demon-
strate that they could see, but they could not discriminate among dif-
ferent objects. For example, the monkeys could press a button when
a light flashed, thus indicating that they could see, but they could not
discriminate among shapes. Later research pinpointed these symp-
toms to disruption of the inferotemporal area (Barlow, 1995).
Other lesion studies in monkeys found cells that appeared to
be specific for face recognition. For example, Rolls and Tovee
(1995) presented monkeys with pictures of faces and other
images. Recording from single cells within the inferotemporal
area, Rolls and Tovee identified a class of cells that responded
strongly to faces, both monkey and human, but barely responded
to nonface stimuli. As we will see in the next section, this finding
is consistent with what we see in the human inferotemporal area
as well (ISLE 5.8).
The Fusiform Face Area
and Face Recognition
One important area within the inferotemporal cortex is known as the
fusiform face area (FFA). The FFA appears to be a specific region in
the brain designed for a specific kind of object recognition, that is, the
recognition of familiar faces (Kanwisher & Dilks, 2013; Kanwisher,
McDermott, & Chun, 1997). The FFA is located on the ventral surface
of the temporal lobe. This area can be seen in Figure 5.27. Although
there is some debate as to how specific the FFA is to face recognition,
most research now suggests that the area is involved in the recognition
of familiar faces after an object has already been perceived as a face
(Liu, Harris, & Kanwisher, 2010). That is, we use the FFA to distin-
guish Aunt Sally’s face from Aunt Mary’s face rather than to identify a strange array of
colors and textures as a face in the first place. Another area of the brain known as the
occipital face area (OFA) appears to be responsible for making the initial identification
of a face as being a face, regardless of its familiarity. That is, the OFA is an area associated
with recognizing faces as distinct from other objects. The OFA is located in the extrastri-
ate cortex and is strongly connected to the FFA (Liu et al., 2010).
Grill-Spector, Knouf, and Kanwisher (2004) examined the role of the FFA in face
recognition using functional magnetic resonance imaging (fMRI) technology. In one
condition of their experiment, they used photographs of the face of the actor Harrison
Ford as their face stimulus. In control conditions, participants saw another face, or
they simply saw a texture gradient with no face. They saw these stimuli for only 50
ms, followed immediately by a masking stimulus, which made it difficult to identify the
stimulus they had just seen. The participants’ task was to identify the briefly presented
photo as Harrison Ford, another face, or nothing at all. The participants did the task
while the fMRI machine was monitoring activity in the FFA.
Grill-Spector et al. (2004) found strong activity in the FFA when participants
identified Harrison Ford in the stimulus. Interestingly, they got strong, though not as
strong, activity in the FFA when the participants thought they saw Ford in the stim-
ulus, but it had actually been a control stimulus. When participants saw the other
face or the texture pattern and identified it as such, there was much less activity
ISLE 5.8
Facial Responses in the Brain
(Inferotemporal Region)
FIGURE 5.27 The Fusiform Face Area (FFA)
and the Occipital Face Area (OFA)
The FFA is an area within the inferotemporal cortex
that is associated with the recognition of familiar
faces in humans. The OFA is located within the
occipital lobe and is associated with recognizing any
face as a face.
Inferotemporal
cortex
Fusiform face
area (FFA)
Occiptal face
area (OFA)
Fusiform face area: an area in
the inferotemporal area of the
temporal lobe that specializes in
recognizing familiar faces
Occipital face area: an area
of the brain in the occipital lobe,
associated with recognizing
faces as distinct from other
objects

139 Chapter 5: Object Perception
in the FFA. Importantly, the response in the
FFA was greatest when the participants recog-
nized Harrison Ford and not the nonfamous
face (Figure 5.28). The FFA shows a selective
response to seeing faces already known rather
than all faces. In summary, this study shows
a strong relation between the recognition of
familiar faces and activity in the FFA.
Prosopagnosia
Damage to the FFA results in a condition known
as prosopagnosia. The term prosopagnosia comes
from two Greek words, one meaning “face” and
one meaning “knowledge.” Prosopagnosia is a
selective deficit in recognizing faces. Prosopagnosia
is a neurological condition in which face recogni-
tion is impaired, but other forms of visual object
recognition are relatively intact. Thus, a patient
with prosopagnosia will have difficulty recog-
nizing particular people but will not have diffi-
culty identifying roses or gloves. It usually occurs
in patients who have had strokes, but there are
some cases of people born with prosopagnosia
(Duchaine & Nakayama, 2006). Patients with
prosopagnosia recognize faces as faces but fail to
recognize specific faces, even of close family mem-
bers. For example, when a prosopagnosic patient
is presented with a photograph of the face of a
family member or a celebrity, the patient will be
able to identify the object as a face but will not be
able to identify the specific person. However, pro-
sopagnosic patients can recognize familiar people
by other cues, such as their voices, their particular
ways of walking, and even distinctive features such
as a big nose or a large facial scar.
One question researchers have debated
is whether prosopagnosia occurs on its own or
whether it is always associated with a more general object agnosia. Most of this research
suggests that, in some cases, a person can have impairment of face recognition but
intact recognition of other objects. For example, Moscovitch and Moscovitch (2000)
compared prosopagnosic patients and object-agnosic patients. Patients with object
agnosia were normal at face recognition but showed deficits in object recognition.
In contrast, the prosopagnosic patients showed deficits in facial recognition but not
object recognition. That is, the prosopagnosic patients, relative to control patients,
were just as able to recognize common objects, such as spoons, pianos, slippers, and
candles. However, these patients could not recognize the faces of people who were
otherwise familiar to them. The object-agnosic patients were able to recognize famil-
iar faces, but they had difficulty recognizing the common objects. This pattern has
now been found in many studies (Busigny & Rossion, 2011).
©
iStockphoto.com
/bloodstone
FIGURE 5.28 The Grill-Spector et al. (2004) Experiment
Participants briefly saw Harrison Ford’s face followed by a mask, or a control
stimulus followed by a mask (a). In (b), we can see the activity in the FFA when
the participant recognized the photo as Harrison Ford and when the participant
did not. Note that the highest response in the FFA is for correct recognition.
Stimulus
Brain activity
measured
Mask
Observer’s
response
Indicate either
(a) “Harrison Ford”
(b) “Another object”
(c) “Nothing”
See either
(a) Harrison Ford
(b) Another person’s face
(c) A random texture
Correct identi�cation
Saw face but
incorrect identi�cation
151050
−0.2
0
0.2
fM
R
I
si
g
n
a
l
Time (sec.)
0.4
Did not see face
(a)
(b)
©
iS
to
ck
ph
ot
o.
co
m
/E
dS
to
ck

140 Sensation and Perception
There is also a very rare condition known as develop-
mental prosopagnosia, which occurs in individuals who
have normal vision and normal social functioning but
show a selective deficit in face recognition (Duchaine &
Nakayama, 2006). Recent fMRI research shows less activity
in the FFA in patients with developmental prosopagnosia,
compared with normal controls, when identifying famil-
iar faces than when viewing other objects (Furl, Garrido,
Dolan, Driver, & Duchaine, 2011). Thus, the evidence with
developmental prosopagnosics also supports the view that
the FFA is an area unique to face recognition.
Other Inferotemporal Cortex
Areas With Specific Object
Recognition Functions
The parahippocampal place area (PPA) is an area within the
inferotemporal cortex that has the specific function of scene
recognition. This area is selectively tuned for the recognition of
spatial landscapes, both indoor and outdoor scenes (Epstein,
2005; Epstein & Kanwisher, 1998). This area is responsive to
photographs of spatial landscapes but shows no additional
activity when images of people are present in those landscapes.
As with the FFA, there is a form of agnosia that accompanies
damage to the PPA. Topographic agnosia involves a deficit in
recognizing spatial landscapes and is related to damage to the
PPA (Mendez & Cherrier, 2003).
Another area in the inferotemporal cortex with a specific function is the extra-
striate body area. The extrastriate body area is activated when its cells view bodies or
body parts but not faces (Downing, Jiang, Shuman, & Kanwisher, 2001). This inter-
esting area is responsible for visual recognition of body areas other than the face. It
may be that other areas of the inferotemporal cortex are also sensitive to particular
stimuli as well. These areas can be seen in Figure 5.29.
Grandmother Cells and Specific
Coding in the Inferotemporal Cortex
An early issue in the history of neuroscience was how specific brain regions are to function
and to particular stimuli. This debate often centered on whether the brain worked as a
whole or was composed of the many separate areas we know it to be composed of now. But
part of the debate concerned where memories were stored in the brain. Thus, for example,
early neuroscientists wanted to know where in the brain was the memory of your cat. When
your cat enters the room, you recognize the cat not just as any cat but as your cat. Clearly,
areas in the inferotemporal cortex are helping you recognize this animal as a cat and as your
good friend Whiskers. The question addressed in this section is whether there is an area of
the brain that is specific to Whiskers but not to your other cat, Fluffball, or any other cat.
In the past, this question was phrased not in terms of cats but rather in terms of another
equally cute category, grandmothers (Figure 5.30). Is there an area in the brain or, more spe-
cifically, in the inferotemporal cortex, that codes for your grandmother but no other person?
Inferotemporal
cortex
Parahippocampal
place
Extrastriate body
Extrastriate
body
Parahippocampal place
area (PPA): an area within
the inferotemporal cortex that
appears to have the specific
function of scene recognition
Topographic agnosia: a
deficit in recognizing spatial
landscapes, related to damage to
the parahippocampal place area
Extrastriate body area: an
area within the inferotemporal
cortex that is activated when its
cells view bodies or body parts
but not faces
FIGURE 5.29 The Parahippocampal Place Area
(PPA) and the Extrastriate Body Area
The PPA is an area within the inferotemporal cortex involved
in scene recognition. The extrastriate body area is associated
with recognizing parts of the body other than the face.

141 Chapter 5: Object Perception
In memory research, this quest has been labeled the
search for the engram. The engram is the specific location
of a specific memory, such as the visual identity of your cat
or your grandmother (Schacter, 2001). Note that the FFA
is responsive to familiar faces but does not necessarily dis-
criminate between the face of your grandmother and that
of your aunt. For many years, the general wisdom was that
such “grandmother” cells would not be found. Identifying
an individual requires the coordination of networks of cells
and communications among them. However, one remark-
able demonstration suggests that grandmother cells may
in fact exist. This study examined famous landmarks and
celebrities rather than grandmothers, but the investigators
found remarkable characteristics of specific object recogni-
tion in inferotemporal neurons (Quiroga et al., 2005).
Quiroga et al. (2005) were able to conduct a very unusual
study because of particular circumstances of their partici-
pants. Their participants were all people about to undergo
elective brain surgery to treat epilepsy. During this procedure, neurosurgeons need
to be very careful about the function of brain regions on which they are about to
operate. As such, as a matter of course, single-cell recordings are taken in some areas
of the brain in these patients. These recordings allow the doctors to determine which
tissue is critical to normal functioning and thus where they need to be extra careful
during surgery. Normally, single-cell recording in live people is not permitted, but in
this case, electrodes were to be implanted in the brains of the patients anyway, and
thus the patients themselves were under no additional risk from Quiroga and col-
leagues’ experimental procedure. Although the results of this study are remarkable,
there has been some controversy over whether these results can be replicated in other
studies.
Quiroga et al. (2005) asked their patients to look at a series of pictures presented
on a computer screen while individual neurons were being monitored. The cells that
were monitored were in the medial temporal lobe, adjacent to but not identical to
the inferotemporal cortex. They found cells that appeared to be specific to individual
people. For example, one cell was selectively responsive to photographs of the bas-
ketball player Kobe Bryant but did not respond to photographs of the actor Mark
Hamill (i.e., Luke Skywalker). Another cell was responsive to photographs of Mark
Hamill but not to Kobe Bryant. The cell responded to Kobe Bryant’s photo with dif-
ferent hairstyles, to a photo of Mr. Bryant and a teammate, and to his printed name.
Different cells in this area appeared to make stronger responses to Mark Hamill as
Luke Skywalker, to other photos of him, and to his name in print. Similarly, Quiroga
et al. found a cell in the medial temporal lobe that was responsive to images of the
leaning tower of Pisa but not to the Golden Gate Bridge, and another cell that was
responsive to the Golden Gate Bridge but not the leaning tower of Pisa. Thus, these
cells are astonishingly like the grandmother cells, responding to a particular person
or landmark from different angles and in different contexts but not to other people
or other famous landmarks (Figure 5.31).
There are some problems, however, with Quiroga and colleagues’ (2005) study,
some of which the researchers themselves have discussed (Quiroga, Reddy, Kreiman,
Koch, & Fried, 2008). First, they did not have much time to test each patient. Thus, they
were not able to test related stimuli. For example, would the “Kobe Bryant” cell also
respond to his fellow Laker star Shaquille O’Neal? Would the “Golden Gate Bridge”
©
iStockphoto.com
/m
ichellegibson
FIGURE 5.30 Grandma and Her Cat
Where in the brain is the representation of the memory of your
grandmother and her favorite cat?

142 Sensation and Perception
cell also respond to a photograph of the San Francisco–Oakland Bay Bridge? These
were questions the investigators were not able to answer, given the time constraints
before the patients’ surgeries. It is also possible, given that the medial temporal lobe
is more involved in memory than visual perception, that the cells were involved in the
retrieval of photograph-related information rather than recognizing specific objects or
people, although such a result would still be surprising. Finally, this study has proved
difficult to replicate, though there has now been at least one conceptual replication
(Gelbard-Sagiv, Mukamel, Harel, Malach, & Fried, 2008). Nonetheless, these studies
show that areas in the temporal lobe may be highly specialized in recognizing partic-
ular objects in the visual world. Thus, Quiroga et al.’s findings are consistent with the
goals of the ventral system—identifying objects in the environment.
TEST YOUR KNOWLEDGE
1. What is prosopagnosia? Is it a form of object agnosia or something different?
2. What is the inferotemporal area? What does it do, and what happens when it is damaged?
5.5 Discuss how we can distinguish human faces from doll
and mannequin faces, and interpret how object perception
research can inform the airport screening process.
FIGURE 5.31 Results of the Quiroga et al. (2005) Study
These data are from single cells within the human temporal cortex. One cell responded to the image of Kobe Bryant but
not to images of other people. Similarly, one cell responded to the image of the Golden Gate Bridge but not to images of
other landmarks.
©
iS
to
ck
ph
ot
o.
co
m
/p
ho
to
qu
es
t7
©
iS
to
ck
ph
ot
o.
co
m
/m
ar
eb
ts
©
iS
to
ck
ph
ot
o.
co
m
/p
ho
to
qu
es
t7
©
iS
to
ck
ph
ot
o.
co
m
/s
te
ve
na
lla
n
©
iS
to
ck
ph
ot
o.
co
m
/E
dS
to
ck
©
iS
to
ck
ph
ot
o.
co
m
/E
dS
to
ck
©
iS
to
ck
ph
ot
o.
co
m
/E
dS
to
ck
©
iS
to
ck
ph
ot
o.
co
m
/E
dS
to
ck

143 Chapter 5: Object Perception
EXPLORATION: Vision and Animacy:
How Do We Tell a Who From a What?
Creepy dolls are almost a cliché in horror movies. Every
couple of years, we see another movie about Chucky, the
crazed doll who terrorizes all who come into contact with
him in the Child’s Play and Chucky movies. Creepy dolls
are a staple of other horror movies, including The Woman
in Black (2012) and one movie aptly named The Creepy
Doll (2011). But dolls are supposed to be playthings for
very young children. How could we give to young chil-
dren objects that produce terror in adults? Store manne-
quins are also often the stuff of horror fantasies, from The
Twilight Zone to Dr. Who. But mannequins are supposed
to inspire us to buy particular clothes. So why do we see
dolls both as cute and harmless toys for our children and
as representations of our deepest fears? And why do seem-
ingly harmless mannequins also bring our fears to the
forefront? Well, we think there is an interesting answer.
Recent research by Thalia Wheatley and her colleagues
at Dartmouth College shed some light on this issue (e.g.,
Looser & Wheatley, 2010).
In Looser and Wheatley’s view, what draws us to dolls
and repels us at the same time is that they have faces
(Figure 5.32). Humans are drawn to see and seek out
faces. Presumably, this was important in human evolu-
tion. Adults need to be able to recognize familiar kin
and distinguish them from potentially dangerous strang-
ers. Young infants must be able to recognize their moth-
ers and distinguish them from other women. Thus, it is
likely that strong evolutionary pressures led to the spe-
cial module for face recognition that exists in the FFA of
the inferotemporal lobe, which allows us to quickly rec-
ognize and identify familiar faces. However, a module
that is designed to seek out faces wherever they are may
seek out faces where none actually exist. It is perhaps
for this reason that we see faces in clouds, mountains,
constellations, and various inanimate objects (Figure
5.33). For many years, for example, the symbol of the
state of New Hampshire was a rock formation on a
mountain that bore a passing resemblance to a human
face. Dolls and mannequins are an interesting case, as
they are deliberately designed to strongly resemble peo-
ple, but they are clearly not people. Dolls have faces, but
they are inanimate objects. Thus, we are drawn to them,
but at the same time, we must recognize that they are not
real human faces. The question that Wheatley’s group
was interested in was what differences might exist in the
perception of real and simulated faces. For a fascinating
take (albeit in a commercial) on our detection of human
faces in odd stimuli, watch this YouTube video: http://
www.youtube.com/watch?v=TQk7Zh-dXCk.
Looser and Wheatley (2010) pointed out that all human
minds have faces, but not all faces have human minds. Dolls
FIGURE 5.32 Dolls Have Faces
A doll has human features but is decidedly not human.
©
iStockphoto.com
/m
itza
FIGURE 5.33 A Smiley Face?
We are tuned to see faces and often see them in objects that do
not really have eyes, noses, and mouths. We often see faces in
clouds, on the sides of mountains, and in tortillas. This is likely
due to an evolutionary disposition to detect human faces.
©
w
anderw
orldim
ages/A
lam
y

144 Sensation and Perception
and mannequins are not the only mindless faces—we have
masks, sculptures, and even photographs and painted por-
traits. So why are dolls and mannequins potentially creepy
but photographs are not? Dolls and mannequins are not
exact duplications of faces (as in photographs) or clearly
iconic representations (as are masks and sculptures). The
better the doll is made, the more closely it resembles a
real human baby or other human figure. Whereas most of
the research we have discussed so far examines how we
detect faces, Wheatley’s research goes a bit further and
asks how we detect faces that have minds attached. When
we see a face, it instantly activates face-sensitive areas of
the brain, such as the FFA. However, animate face recog-
nition requires an additional step: to associate the face
with a moving, living, animate being. Thus, we must make
an additional step to go from face to mind.
This is supported by Wheatley’s research. Wheatley,
Weinberg, Looser, Moran, and Hajcak (2011) used
electroencephalographic (EEG) technology to look at
responses to photographs of real faces and photographs
of dolls and mannequins. The EEG responses were aver-
aged to look at characteristic wave patterns, known as
event-related potentials, in response to these face stimuli.
Wheatley et al. found that they could identify an event-
related potential wave that occurred after recognizing
faces, real or inanimate, but not other stimuli, presum-
ably emanating from the FFA. When dolls and manne-
quins were presented to participants, the face response
on the EEG decreased about 200 ms after presentation.
However, for real faces, the EEG response continued.
Wheatley’s team could therefore tell what kind of face,
animate or inanimate, a person had
examined by looking at the EEG
response. Thus, human visual sys-
tems can quickly distinguish between
human faces and inanimate replicas.
In the Looser and Wheatley (2010)
experiment, they combined the
images of real faces and inani-
mate faces, as can be seen in Figure
5.34. They found images of statues
and dolls that matched the faces
of actual people and, using a com-
puter program, morphed the images
together. Looser and Wheatley then
varied the level at which each image
contained more human or more
doll in the face, with some being all
doll and some being all human, as
well as various in-between levels.
Participants then made a number of
different judgments concerning the
faces, including alive or not alive,
realistic or unrealistic, has a mind
or does not have a mind, and able
to feel pain or not able to feel pain.
Regardless of the particular judg-
ment, Looser and Wheatley found a
tipping point at about 65% human.
That is, if the combined image was
at 65% human face or higher (i.e.,
35% doll), it was judged to be alive,
to be realistic, to have a mind, and to
FIGURE 5.34 Transformation From Real Baby to Doll
Consider the middle two faces. Are they of a doll or a person?
FIGURE 5.35 Results of Looser and Wheatley’s (2010) Study of Animacy
These data show participants’ perceptions of whether the figures were animate or not.
The colored lines indicate the ratings for attributes of animacy, such as animacy itself,
having a mind, the ability to plan, and the ability to feel pain. From Looser & Wheatley
(2010), Figure 2, p. 1856.
De�nitely
has
Attribute
De�nitely
lacks
S
u
b
je
ct
s’
p
e
rc
e
p
ti
o
n
s
o
f
a
tt
ri
b
u
te
Actual physical difference
PSE
1.0.9.8.7.6.5.4.3.2.1.0
.0
.1
.2
.3
.4
.5
.6
.7
.8
Animacy
Plan
Pain
.9
1.0
*
Mind

145 Chapter 5: Object Perception
APPLICATION: The Science of Airport Screening
Most of you are probably familiar with the chaotic
scenes at crowded modern airports. As we walk to our
gates at the airport, we dodge people with large roll-
ing suitcases, strollers, and wheelchairs, and we also
now dodge the people who are so trustworthy of others
that they feel comfortable texting and walking in large
crowds. Equally ubiquitous are the inevitable lines that
form as we go from unsecured areas to secured areas
in the airport. At airport security, all travelers and their
bags must go through machines designed to help agents
detect objects that travelers are not allowed to bring
on airplanes. These objects may vary from seemingly
innocuous objects such as jars of jam to much more
dangerous objects such as knives, box cutters, and guns.
Although these machines are very adept at looking into
your belongings, there still must be a pair of human eyes
watching the ongoing stream of bags. These security offi-
cials are looking for something very uncommon, as most
of us know better than to pack forbidden objects for
air travel. Thus, the dangerous object, such as a knife,
is among a tremendously greater number of innocuous
objects. Part of the job of such a security official (TSA
agents, in the United States) is the ability to recognize the
dangerous objects among all of the harmless ones in real
time under rushed and degraded conditions.
The general study of object perception and applied stud-
ies on detecting dangerous objects in X-ray displays have
tremendous importance for our system of deterring crime
in the air. These studies should have an influence on how
feel pain. At less than 65%, the animacy ratings tailed
off quickly (Figure 5.35). Looser and Wheatley found it
significant that this tipping point was higher than 50%.
They think it is important that we have a high criterion
for recognizing animacy in a human face.
Looser and Wheatley (2010) compared their results with
earlier social-psychological data on object perception
that showed that, in many situations, we find facelike
inanimate objects to be likable (e.g., a rock with eyes
drawn on it), but we are often suspicious of or revolted
by clearly inanimate objects that look too human (e.g.,
mannequins, dolls). It may be for this reason that many
children’s cartoons use anthropomorphic animals or
caricature images of children to portray their characters
rather than more lifelike representations.
In another experiment, Looser and Wheatley (2010)
examined individual features of faces, including the eye,
the mouth, the nose, and the skin, and then compared
the judgments of animacy from just these facial features
with those from whole faces. Overall, they found lower
animacy judgments for parts of a face than for a whole
face, but they also found that animate-looking eyes were
the biggest predictor of animacy. Seeing an eye by itself
allowed participants to judge animacy in a way in which
mouths and noses did not. In evaluating inanimate faces,
it may be that the other features of the face activate the
brain’s face detectors, but the lack of animate eyes may
give a doll or mannequin its “creepy” look (Figure 5.36).
To summarize this research, human visual systems are
extremely sensitive to stimuli that look like human faces.
This means that we may often get “false alarms,” that
is, stimuli that look like human faces but are actually
not. We can quickly reject faces in stones and clouds as
not being really human, but dolls and mannequins give
us pause. In most cases, we reject dolls as being human,
but perhaps every so often, they tweak our “mind”
detectors. This causes us to wonder, if only for a second,
if they are watching us and observing our actions. This
leads to a creepy feeling that the old unused doll gath-
ering dust in the closet is really plotting its revenge on
us. Thus, the creepy feeling we get from dolls is really
the activation of face areas in the brain, the activation
of “mind” areas in the brain, but the knowledge that
the doll does not have a mind. This causes us to worry
about what might lurk behind that face, and danger and
creepiness result.
FIGURE 5.36 The Perception of Creepiness
Is this creepy or what?
©
iStockphoto.com
/sqback

146 Sensation and Perception
FIGURE 5.37 Airport Security Checkpoint
Ed
uc
at
io
n
Im
ag
es
/U
ni
ve
rs
al
Im
ag
es
G
ro
up
/G
et
ty
Im
ag
es
screeners look for dangerous objects, how those screeners
are trained to look for dangerous objects, and how the
technology can be modified in the future to make it easier
for TSA agents to do their job effectively. Let’s consider a
few of these studies and their recommendations.
Airport screeners are looking for any number of contra-
band objects, from concealed firearms to oversized tooth-
paste. While on duty, each screener is looking for all of
these objects simultaneously. Research suggests, however,
that screeners would be better at detecting forbidden
objects if they were only scanning for one type of object.
That is, one screener should be looking just for guns, while
another looks just for knives. The research shows that as
the number of objects for which a screener is looking goes
up, the harder it is to detect any of these objects individu-
ally (Godwin et al., 2010). That is, if a screener is looking
for just one type of object (e.g., firearms), he is more likely
to detect a hidden firearm than if he is also looking for
knives and oversized liquids. This problem with searching
for multiple types of objects is known as the dual-target
cost. The dual-target cost means that as the number of
objects searched for increases, the likelihood of detecting
one of those objects decreases (Godwin et al., 2010; Smith,
Redford, Washburn, & Taglialatela, 2005). This dual-target
cost shows up in the accuracy of detecting objects as well
as the time it takes to recognize a to-be-rejected item.
It occurs in the lab for random objects, and it occurs in
simulated screening environments for banned weapons
(Godwin et al., 2010).
Eye-movement tracking shows that the dual-target costs
result from ineffective guidance during the visual search
process (Menneer, Stroud, Donnelly, & Rayner, 2008). In
this research, when participants are looking only for a sin-
gle type of target, they look at any object in the array that
resembles the target, thus allowing them to pick up on slight
deviations in the basic pattern (e.g., a handgun with a short
barrel but a large magazine). However, when participants
are looking for two types of objects, they tend to perse-
verate on objects that resemble neither target. This makes
it not only more difficult to pick out objects that deviate
from the basic pattern but also more time-consuming to
find objects that are typical of the category. As such, time is
wasted examining objects not part of the screening process
(Menneer et al., 2008). This wasted time may be inconse-
quential in a laboratory task, but when you have a huge line
forming at Security in an airport, it may encourage screeners
to rush through without adequately checking all items. This
finding has been shown not only in many laboratory studies
with more limited realism to airport screening but also in a
situation that directly mimics what screeners do at security
at airports (Godwin et al., 2010).
Can agents be trained to more accurately detect banned
objects? The research here is mixed. Some research
points to improvements with training, but other
research does not (see Nodine, Krupinski, & Kundel,
2002). It turns out it depends on what aspects of the
task the training is directed toward. McCarley, Kramer,
Ca
m
az
in
e
Sc
ot
t/
Sc
ie
nc
e
So
ur
ce
/G
et
ty
Im
ag
es(a) (b)
Dual target-cost: as the number of objects searched for
increases, the likelihood of detecting one of those objects
decreases

147 Chapter 5: Object Perception
Wickens, Vidoni, and Boot (2004) trained participants
in a simulated airport screening task. Participants had
to examine X-rays and look for knives in the suitcases.
McCarley et al. found that people got better and were
able to do the task faster as they practiced more and
more. However, their analysis of their participants’ eye
movements suggested that what the participants were
learning was how to get better at detecting specific
objects, rather than their ability to quickly move their
eyes and scan across the bag. In addition, the improve-
ment was limited to the specific objects on which the
participants were trained. When they were then asked
to look for different kinds of knives, the improvement
based on practice was lost (McCarley et al., 2004).
Other research shows that even after training, the
dual-target costs remain (Godwin, Houpt, Walenchok,
Hout, & Goldinger, 2015).
This research suggests that a better model of screen-
ing might be to train each TSA agent for a specific type
of banned object. Thus, while one agent is looking for
knives, another is looking for handguns, and another is
looking for liquid explosives. Given that each agent is
looking for only one class of object, they will benefit
from training and practice and will be better at detecting
that danger, leaving all of us safer when we travel.
CHAPTER SUMMARY
5.1
Interpret the computational difficulties the visual sys-
tem must overcome in recognizing objects as them-
selves under a multitude of situations and angles.
Object perception concerns our ability to visually identify the
world around us, from socks to bird feeders to our Aunt Sally.
Object agnosia is an acquired deficit in identifying and rec-
ognizing objects even though vision remains intact.
5.2
Assess the difference between top-down pro-
cessing and bottom-up processing and how they
affect object perception.
Object perception requires both bottom-up processing,
from stimulus to perception, and top-down processing, from
knowledge to perception, in order to function efficiently.
Object perception requires a memory representation system
that allows us to recognize familiar objects when we see
them.
5.3 Describe the gestalt laws of perceptual grouping.
Perceptual organization is the process by which multi-
ple objects in the environment are grouped, allowing us to
identify those objects in complex scenes. Two important
processes in perceptual organization are grouping and
segregation. Gestalt psychology argues that what we see
is greater than the sum of its individual parts and, as such,
emphasizes the rules that govern how objects in scenes are
determined from background noise. One set of these rules
concerns figure–ground organization, which refers to the
inference viewers must make as to which part of an image is
in front and which part of an image is in the background of a
particular scene. Gestalt psychologists also emphasized the
laws of perceptual grouping, which allow us to infer which
parts of an image go with which other parts. One such illus-
tration of these principles comes from illusory contours, in
which we see figures present that are not part of the physi-
cal stimulus.
5.4
Explain why the ventral pathway is critical for
object perception.
Object perception occurs along the ventral pathway in
the brain, which leads from V1 in the occipital lobe to the
inferotemporal area of the temporal lobe. V4 is an area in the
extrastriate cortex involved in shape perception. The infero-
temporal area is the area of the temporal lobe involved in
object perception. It receives input from V4 and other areas
in the occipital lobe. Part of the inferotemporal lobe is known
as the fusiform face area (FFA). The FFA is an area in the
inferotemporal cortex of the temporal lobe that specializes in
recognizing familiar faces. There is also a region known as
the occipital face area (OFA), which is an area of the brain in
the occipital lobe that is associated with recognizing faces
as distinct from other objects. Damage to the FFA results in a
condition known as prosopagnosia. The term prosopagnosia
comes from two ancient Greek words, one meaning “face”
and one meaning “knowledge.” Prosopagnosia is a selec-
tive deficit in recognizing faces. The parahippocampal place
area (PPA) is an area within the inferotemporal cortex that
appears to have the specific function of scene recognition.
Topographic agnosia involves a deficit in recognizing spa-
tial landscapes, related to damage to the PPA. Quiroga et al.

Sensation and Perception148
(2005) conducted a study that found cells in the temporal lobe
that responded to specific people, regardless of whether a
photograph or a name was presented. Some people regard
this study as evidence for neural areas that code for very
specific information about individual objects or people.
5.5
Discuss how we can distinguish human faces
from doll and mannequin faces, and interpret how
object perception research can inform the airport
screening process.
When we see a face, it instantly activates face-sensitive areas
of the brain, such as the FFA. However, animate face recog-
nition requires an additional step: to associate the face with
a moving, living, animate being. Research by Wheatley and
her colleagues shows that faces are recognized rapidly, but
an extra step exists to distinguish real human faces from fac-
similes, such as dolls. Human agents must inspect bags as
they go through airport screening. This is a task very much
dependent on people’s ability to recognize specific objects,
in this case, dangerous ones, such as guns and knives.
Research shows a dual-target cost, which means screen-
ers are slower and less accurate when they are looking for
multiple objects. Moreover, improvements in training occur
when screeners are scanning for only one class of objects.
Looking for multiple objects slows people down. Thus, the
research suggests that specific agents train for the detec-
tion of only one class of dangerous object.
REVIEW QUESTIONS
1. What is object agnosia? What areas of the brain
likely cause it, and what are its symptoms?
2. Define object perception. What are some of the
obstacles programmers must overcome to design a
computer system with object perception?
3. What is the difference between top-down and bot-
tom-up processing?
4. What is meant by perceptual organization? What is
the difference between grouping and segregation?
5. Who were the gestalt psychologists? How did they
contribute to our understanding of figure–ground
perception?
6. What is edge completion? How does it account for
illusory contours?
7. Describe the neural pathway that leads from V1
to the temporal lobe that is responsible for object
perception.
8. What is the difference between the FFA and the
OFA in perceiving faces? Which area is damaged in
prosopagnosia?
9. Several experiments have demonstrated the role of
the FFA in face perception. Describe one of these
experiments and how it confirms this area’s function.
10. Why is judging animacy in a face important? Why do
dolls and mannequins elicit “creepy” responses, but
pictures of real faces do not? What difference do we
see in EEG scans when people view photographs of
real faces and those of dolls?
PONDER FURTHER
1. Do you think recognizing words and letters follows
similar principles to the ones outlined here for object
perception? How might you test, in a behavioral exper-
iment, whether recognizing letters follows the same
gestalt principles as recognizing routine objects?
2. Faces are clearly important stimuli to human beings.
What other kinds of visual stimuli are as important
as faces? Might there be specific areas of the brain
for these classes of objects? How might you test for
these areas?
KEY TERMS
Bottom-up processing, 126
Dual-target cost, 146
Edge completion, 134
Extrastriate body area, 140
Figure–ground organization, 129
Fusiform face area, 138

Chapter 5: Object Perception 149
Geons, 136
Grouping, 128
Illusory contours, 135
Inferotemporal area, 137
Law of common fate, 134
Law of good continuation, 133
Law of proximity, 133
Law of similarity, 133
Law of symmetry, 133
Occipital face area, 138
Parahippocampal place area (PPA), 140
Perceptual organization, 128
Recognition, 127
Recognition by components, 136
Representation, 127
Segregation, 128
Top-down processing, 126
Topographic agnosia, 140
V4, 137
Viewpoint invariance, 136
Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e
SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review,
study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning.
Learning Objectives Digital Resources
5.1 Interpret the computational difficulties the visual system
must overcome in recognizing objects as themselves under a
multitude of situations and angles.
From Fragments to Geometric Shape: Changes in Visual
Object Recognition Between 18 and 24 Months
Visual Agnosia
Object Agnosia
5.2 Assess the difference between top-down processing and
bottom-up processing and how they affect object perception.
Top-Down Visual Processing
5.3 Describe the gestalt laws of perceptual grouping. A Perceptually Completed Whole Is Less Than the Sum of
Its Parts
5.4 Explain why the ventral pathway is critical for object
perception.
Object Recognition: Insights From Advances in fMRI Methods
Prosopagnosia
Face Blindness, Part 1
5.5 Discuss how we can distinguish human faces from doll and
mannequin faces, and interpret how object perception research
can inform the airport screening process.
Thalia Wheatley: How the Brain Perceives Other Minds

6Color Perception
David Parker/Science Source
LEARNING OBJECTIVES
6.1 Examine the relation of the wavelength of light to perceived color.
6.2
Diagram the perceptual principles of hue, saturation, and
brightness and the role they play in our perception of color.
6.3
Formulate the idea of a metamer in additive and subtractive color
mixing in terms of what we learn from color-matching experiments.
6.4 Examine the role of having three cone classes for our perception of color.
6.5
Describe the trichromatic theory of color vision and its
relationship to the three classes of cones.
6.6
Illustrate the opponent-process theory of color vision focusing on what this
theory can explain about color vision that trichromatic theory cannot.
6.7 Explain what we know about color vision in infancy
and how it changes as we age.
6.8
Diagram the different types of color deficiency and explain why this
term is, almost always, a better term than color blindness.
6.9 Assess the term constancy and how it applies to color vision.
INTRODUCTION
If you live in South Florida, seeing rainbows is a frequent occurrence. In the summer,
rainbows are quite common after a late afternoon thunderstorm. Driving along Route
1 in the Florida Keys in the summer, one can often see layer upon layer of rainbows
as one heads east in the late afternoon (Figure 6.1). Many people find rainbows to be
beautiful, partly because they appear to be pure color—just color, not attached to a par-
ticular object, seemingly magically suspended in the air. The physical explanation for
rainbows, millions of water droplets acting as prisms, just makes the phenomenon more
mysterious. We know it has to be true, but for many, the physical explanation just does
not connect with their perceptual experience. When most of us look at rainbows, we see
seven distinct bands: red, orange, yellow, green, blue, indigo (a dark blue), and violet
ISLE EXERCISES
6.1 Different Types of
White Lights
6.2 Dimensions of Color
6.3 Newton’s Prism
Experiment
6.4 Color Mixing
6.5 Color-Matching
Experiment: Metameric
Matches
6.6 Trichromatic Theory
and Cone Responses
6.7 Univariance and
Color Matching in
Monochromat or During
Scotopic Vision
6.8 Color Aftereffect
6.9 Simultaneous Color
Contrast
6.10 Hue Cancellation
6.11 Single- and Double-
Opponent Cells
6.12 Color-Deficiency Tests
6.13 Rod Monochromat
Vision
6.14 Dichromacy
6.15 Illumination and
Color Constancy
6.16 Lightness Constancy
6.17 Gelb Effect
6.18 Color Camouflage
and Dichromacy

152 Sensation and Perception
(a reddish blue resembling purple). However, a
rainbow is actually a continuous band of wave-
lengths—the seven bands are a function of our
eyes and brains. As we will see in this chapter,
colors are usually properties of objects, but in a
rainbow, we do not see the raindrops that create
the rainbows.
So, do all people and animals see color the
same way? Many of you may have heard that
your cat is color-blind. This is not quite true.
Cats do have some ability to see in color but
much less than we do (Buzás et al., 2013). Cats
are actually “color deficient,” meaning that
they have a two-cone color system, compared
with our three-cone system. Cats also have
fewer cones in total than humans, but they
can see colors. Cats’ eyes are better adapted
than ours for nocturnal vision, and their eyes
are more specialized for detecting low levels of
light at night. (Review the Exploration section
of Chapter 3.) Nonetheless, because cats are
active during the day, they also have some color vision. Dogs are also color deficient
relative to us, but they do see in color as well. It turns out that their retinae resemble
those of color-deficient people, in having a two-cone system. You probably know a
person who is red–green color-blind (better called, “color deficient”). Whereas you
see a green light or a red light at a traffic signal, he sees both the green and the red
lights as the same color. Luckily, he knows that the red light is always above the green
light, so red–green color deficiency is not a risk for driving. Are there animals with
better color vision than ours? You may have heard that goldfish have a four-cone
system, compared with our three-cone system. On the basis of research on goldfish,
it is known that they can see color differences that look the same to us (Neumeyer &
Arnold, 1989). So, some animals see fewer colors and some see more. What about
other people?
Interestingly, it turns out that some people may also have functional four-cone
systems that allow them to see more colors than the more common people with three
cones do. Gabriele Jordan, a researcher at the University of Newcastle in the United
Kingdom, recently identified a woman with a functional four-cone system, who seem-
ingly can see different colors when most of us would just see one color (Greenwood,
2012; Jordan, Deeb, Bosten, & Mollon, 2010). However, Jordan and her colleagues
think functional four-cone vision is extremely rare in people. In their study, they looked
at the genetics that cause some women to have four-cone systems and then tested
behaviorally whether women with four-cone systems could actually discriminate more
colors. Because many of the genes for color vision are on the X chromosome, men are
more likely to be color deficient and less likely to have the extra cone system. Jordan
et al. examined many people (all women) who have a genetic mutation that gives them
what looks like an extra cone system in their retinae. The question was whether this
would be expressed in terms of seeing more different colors than trichromats do. Only
one of these women actually was able to discriminate colors that looked identical to
the other participants or to controls with three cones. Jordan et al. estimated that this
woman may see millions more hues than other people do. Current research by Jordan’s
©
R
EU
TE
RS
/M
ar
c
Se
ro
ta
FIGURE 6.1 Rainbow
A rainbow is a continuous band of wavelengths that we see as seven colors blending
into one another. In addition to being a beautiful sight, a rainbow can tell us a bit
about color perception.

153 Chapter 6: Color Perception
group aims to understand what this woman sees, how it differs from typical color
vision, and whether it really gives her abilities the rest of us do not possess. If color
vision varies across animals and even people, how does color vision happen and what
allows for these variations?
WAVELENGTHS OF LIGHT AND COLOR
6.1 Examine the relation of the wavelength of light to perceived color.
In Chapter 3, we introduced the idea of the wavelength of light and its relation to
the perception of color. Humans can see wavelengths of light that vary between 400
and 700 nm, often known as the visual spectrum (or visible spectrum). Within the
visual spectrum, we see different wavelengths as different colors. It is important to
keep in mind that the physical stimulus is the wavelength of the light entering our
eye. Color is a perceptual attribute, or our response to this physical feature of light.
The term wavelength refers to the distance between two peaks of light energy. It
is the inverse of frequency. As frequency increases, wavelength decreases. In color
vision, the wavelength of light is usually referred to, and as we will see, with sound
waves, we tend to use frequency instead of wavelength (just to keep students on
their toes).
Sunlight is a mix of many different wavelengths, which blend together to form
white light. Because the wavelengths are bunched together within our acuity limits,
we cannot discriminate the different wavelengths in white light. But because sunlight
is composed of so many wavelengths, raindrops can diffract (i.e., spread out) the light
into the multiple wavelengths that make up the rainbow. The raindrops act as a prism
and separate the many wavelengths that are mixed together. Once they are separated,
we can see the individual bands of wavelengths as colors. Natural sunlight, however,
varies in the distribution of wavelengths during the course of a day. At noon, there is
more blue light than in the morning, and evening is well known for its lingering red
light. Many artificial sources of light tend to be mixes of multiple wavelengths that can
also be classified as white light. Normal incandescent light bulbs tend to emit more
long-wavelength light than short-wavelength light. This gives incandescent light bulbs
their characteristic yellow color. Fluorescent light bulbs tend to have the opposite pat-
tern, giving them their slightly blue hue (Figure 6.2 and ISLE 6.1). All of these are vari-
ants of white light, that is, heterochromatic light consisting of many wavelengths. In
contrast, monochromatic light, which can be produced by special light bulbs, is light
of only one wavelength or a very narrow band of wavelengths. When monochromatic
light reflects off of a white surface, we see that surface as the color associated with that
wavelength.
Another important feature in the perception of color is the spectral reflectance of objects
in the world. Spectral reflectance is the ratio of light reflected by an object at each wave-
length. This means that every object has particular characteristics that permit it to absorb
some wavelengths of light and reflect other wavelengths of light. Thus, white clothes reflect
wavelengths equally, whereas blue jeans absorb most wavelengths but reflect light at about
450 nm, which we perceive as blue. Thus, the color of any object is determined by what
wavelengths it reflects the most. Leaves are green because their surfaces absorb most light
but reflect those wavelengths we see as green. Carrots are orange because they absorb
most wavelengths but reflect light at about 650 nm, which we perceive as orange. Figure 6.3
Visual spectrum (visible
spectrum): the band of
wavelengths from 400 to 700
nm that people with the most
common classes of cones
vision can detect
Heterochromatic light:
white light, consisting of many
wavelengths
Monochromatic light: light
consisting of one wavelength
Spectral reflectance: the
ratio of light reflected by an
object at each wavelength
ISLE 6.1
Different Types of White Lights

154 Sensation and Perception
FIGURE 6.2 The
Distribution of Wavelengths
in Natural and Artificial Light
(a) The top part of the figure shows
the visible spectrum of light on
a continuum of electromagnetic
radiation. The figure makes it clear
that visible light is just a small range
on this continuum. The bottom part
of the graph shows the distribution
of wavelengths across natural
sunlight, incandescent light bulbs,
and fluorescent bulbs. Note the
interesting peaks in the fluorescent
bulbs, which are very strong in the
blue and green range, whereas
incandescent light bulbs show a
gradual increase, peaking in the
longer red wavelengths. (b) In
this figure, two different types of
artificial illumination are visible, both
considered white. Over the museum
and lounge, the light source has
more long wavelengths, and over the
hallway, the light source has more
short wavelengths. Because the two
light sources are next to each other,
you can see the differences in the
whites more easily.
Visible Spectrum
Gamma rays
10–11 10–9 10–7 10–5 10–3 10–1 101 103 105 107 109 1011 1013 1015 1017
Radio waves
Spectrum of
electromagnetic
radiation
x-
rays
UV
rays
Infrared
radiation
Wavelength (nm)
Micro-
waves
Monochromatic light
Wavelength (nm)
400
0
1
2
3
4
5
6
500 600 700
R
e
la
ti
v
e
i
n
te
n
si
ty
Sunlight
Fluorescent bulb
Incandescent bulb
shows objects in their characteristic color and different colors to
illustrate how reflectance is used to determine an object’s color
and how we rely on that characteristic color.
Surfaces that reflect all light equally can be said to be ach-
romatic (which means “without color”) because we do not see
them as containing color. Achromatic surfaces are judged to be
white to gray to black (Figure 6.4). With such surfaces, what
matters is the proportion of ambient light they reflect. Surfaces
that reflect most of the light that hits them will be seen as white
surfaces (e.g., 90% or higher reflectance). Surfaces that reflect
some but not all light will be seen as gray (e.g., 50% reflectance),
and those that absorb most light will be seen as black (10% or
lower reflectance). An important property of our visual system
is that we respond to the percentage reflected rather than the
FIGURE 6.3 Familiar Colors
Objects with their familiar characteristic colors, and the same
objects but with odd colors. Objects get their characteristic colors
because they reflect light at particular wavelengths. Ripe bananas
absorb most light but reflect light at about 550 nm, giving them their
yellow color.
(a)
(b)

155 Chapter 6: Color Perception
total amount reflected. This allows us to see surfaces as the same
level of grayness despite changes in the total amount of light. Thus,
we still see a gray cat’s fur as gray in bright sunlight or by a dim
lamp in the evening, even though the fur is reflecting much more
total light in broad daylight. With this background, we can now
turn to perceptual features of color.
TEST YOUR KNOWLEDGE
1. How do the colors of the rainbow map onto wavelength as
measured in nanometers?
2. Describe the difference between monochromatic and
heterochromatic lights.
HUE, SATURATION, LIGHTNESS,
AND BRIGHTNESS
6.2
Diagram the perceptual principles of hue, saturation, and brightness
and the role they play in our perception of color.
The perception of color is often described by referring to three dimensions of the
color experience: hue, saturation, and brightness. It is recommended that you fol-
low along while reading this section with having ISLE 6.2 open to help you expe-
rience these dimensions. We start with hue. Hue refers to the color quality of the
light and corresponds to the color names that we use, such as orange, purple, green,
indigo, yellow, cyan, aquamarine, and so on. In fact, hue is the quality of color. A
quality is a value that changes, but it does not make the value larger or smaller.
For example, when intensity (the amount of light present) changes, it gets larger or
smaller. But when hue or color changes, it does not make sense to say, for example,
that red has more or less hue than green. This is because color is a quality, not an
amount. We see red and green as distinct units, even though some greens may seem
darker or lighter, but this corresponds to the dimension of brightness (discussed
later in this section). But certainly, red is very different from green. Many hues
have direct correspondence to particular wavelengths, such as red (650 nm) and
green (550 nm). Colors that are associated with particular wavelengths are called
monochromatic colors, which include the basic colors or spectral colors, such as
red, green, orange, yellow, and blue. There are also colors known as nonspectral,
which are made of combinations of more than one monochromatic color. These are
colors such as purple, brown, silver, and gold. Brown, for example, is a yellow or
orange spectral color with very low saturation, that is, mixed with black. Purple is
usually a mix of red and blue, though it can be combined in other ways as well.
Saturation refers to the purity of light. The more saturated the stimulus is, the stronger
the color experience will be, and the less saturated the stimulus is, the more it appears
white or gray or black, that is, achromatic. The classic example of saturation differences
concerns the continuum from red to pink. Pink is a combination of red light and white
light. The more that white light is added, the less “red” the pink is. A pastel pink may
contain just a bit of red light along with a lot of white light. Eventually, the red may be so
overwhelmed by the white that we barely notice the pink at all. A solid blue becomes less
saturated as we add more white, eventually becoming the “baby blue” we might paint a
baby’s room’s walls (Figure 6.5).
FIGURE 6.4 Achromatic Lightness
When objects reflect all light wavelengths equally, they are
said to be achromatic. The three squares depicted here do
not show any differences in the reflectivity of wavelengths.
However, the white square reflects most light that shines on it,
the gray square reflects about half of the light that shines on it,
and the black square absorbs most of the light that shines on it.
ISLE 6.2
Dimensions of Color
Hue: the color quality of
light, corresponding to the
color names we use, such as
orange, green, indigo, and
cyan; hue is the quality of
color
Quality: a value that changes
but does not make the value
larger or smaller
Saturation: the purity of light
FIGURE 6.5 Saturation
On the left side are hues with strong
saturation. We see these colors as very
strong. On the right, we see hues with less
saturation. The red becomes pink, and the
blue becomes baby blue. The figures at the
right have more white light mixed with the
red or blue light in the shape.

156 Sensation and Perception
Hue and saturation can be represented as a color circle
(Figure 6.6). Along the perimeter of the circle, we find hues,
such as red, orange, yellow, green, blue, indigo, and violet.
As we head toward the center of the circle, we get less and
less saturated colors. Thus, a deep red and a deep blue exist
on the perimeter, but pink and baby blue exist toward the
center. You can explore an interactive illustration of these
principles on ISLE 6.2.
Brightness is related to the amount of light present and
is our experience of the intensity of this light. The brighter
an object is, the easier it is to see and the more notice-
able the colors are. Brightness is the psychological dimen-
sion that goes vertically through the color circle, now a
color solid (Figure 6.7). Brightness does have a relation to
color—it is easier to see color at higher brightness values.
Brightness is distinguished here from lightness. Lightness
refers to the perceived intensity that relates to the amount
of light that gets reflected by a surface. This is a different
property than the amount of light present. Brightness usu-
ally applies to colors, whereas lightness usually refers to
the white-gray-black continuum.
TEST YOUR KNOWLEDGE
1. How is hue related to wavelength?
2. Describe the difference between brightness and intensity.
ADDITIVE AND
SUBTRACTIVE COLOR MIXING
6.3
Formulate the idea of a metamer in additive and subtractive color
mixing in terms of what we learn from color-matching experiments.
Color has an odd characteristic that is suggested by the color circle. Colors can be
mixed. That may not seem odd, but that is because you have been doing it for so
long—probably first during preschool art sessions. Usually, children in kindergarten
have a pretty good intuitive sense about how to mix paints. But it is a relatively unique
feature of color within the domain of visual perception, and the science is quite com-
plex. To understand this statement, consider the following questions: Can you try
to mix forms or motions? Can you mix size and depth? Most of our sensations and
perceptions do not mix in the ways that colors do. Isaac Newton’s classic experiments
with prisms illustrate the profundity of the idea that colors mix (Figure 6.8). You can
explore an interactive illustration of this principle on ISLE 6.3.
There are two main types of color mixing: subtractive color mixing and additive color
mixing. Additive color mixing is the creation of a new color by a process that adds one
set of wavelengths to another set of wavelengths (remember spatial summation). Additive
color mixing is what happens when lights of different wavelengths are mixed. When we add
all of the different wavelengths of sunlight, we see white light rather than many individual
colors. This is called additive because all of the wavelengths still reach our eyes. It is the
combination of different wavelengths that creates the diversity of colors. Subtractive color
FIGURE 6.6 The Color Circle
The color circle is a two-dimensional representation of how hue and
saturation interact. Hue is represented around the perimeter, whereas
saturation is represented as distance from the center (along the radius
of the circle). The most saturated hues are along the perimeter, and
saturation decreases as one moves toward the center.
Green
(550 nm)
Yellow
(600 nm)
Hue
Red
(650 nm)
Cyan
(500 nm)
Saturation
No
ns
pe
ctr
al
pu
rpl
es
Blue
(450 nm)
FIGURE 6.7
The Color Solid
The color solid is a three-dimensional
representation of the relations
between brightness, saturation, and
hue. As in the color circle, hue is
represented around the perimeter,
whereas saturation is represented
as distance from the center (along
the radius of the circle). The new
dimension, brightness, is represented
by the vertical dimension.
Hue
Saturation
Brightness
ISLE 6.3
Newton’s Prism Experiment

157 Chapter 6: Color Perception
mixing is the creation of a new color by the removal
of wavelengths from a light with a broad spectrum of
wavelengths. Subtractive color mixing occurs when we
mix paints, dyes, or pigments. When we mix paints,
both paints still absorb all of the wavelengths they did
previously, so what we are left with is only the wave-
lengths that both paints reflect. This is called subtrac-
tive mixing because when the paints mix, wavelengths
are deleted from what we see because each paint will
absorb some wavelengths that the other paint reflects,
thus leaving us with a smaller number of wavelengths
remaining afterward. The easy way to remember the dif-
ference between additive and subtractive color mixing
is that additive color mixing is what happens when we
mix lights of different colors, whereas subtractive color
mixing occurs when we mix paints or other colored
materials.
In the theater, lighting technicians use colored lights
to create additive color mixing to illuminate the stage.
When we look overhead at the lights, we may see a vari-
ety of colored lights projecting onto the stage, although
the mix of these lights may create a different color than
any present in the bank of lights overhead. Professional
painters (and preschool children) can create a larger
of palette of colors by mixing different paints, thereby
using subtractive color mixing. But if you do not know
what you are doing when mixing paints, the result very
often looks muddy and unappealing. We now look more
closely at the science of each form of color mixing.
Additive Color Mixing
(Mixing Lights)
Additive color mixing occurs when lights of different
wavelengths are mixed. This is what occurs in televisions. All of the colors one sees on a
television screen are the result of the mix of three different light sources built into the tele-
vision and controlled at different intensities. The same is true for our computer screens.
What causes us to see colors such as chartreuse on our computer screens is a carefully bal-
anced mix of three lights. When we shine equally bright green and red lights on the screen,
the lights add and we see yellow. However, if the red light is brighter than the green light,
you will most likely perceive an orange hue. This is how televisions can create so many
colors using just three dots. The principle of additive color mixing is illustrated in Figure
6.9. You can see a demonstration of additive color mixing on ISLE 6.4.
In Figure 6.9, you can see the color circle. It can be used to predict perceived colors
in additive mixtures. The center of the circle represents the point at which any additive
mixture contains equal amounts of the different wavelengths present. This center point is
perceived as a gray. If you draw a line between two colors on the circle, say yellow and
blue, the perceived color will fall on that line, depending on the relative intensities of the
two lights. When you mix three lights, the color form is a function of a triangle between
the three primary lights. Exactly what hue is perceived within that triangle is a function
of the relative intensity of the three primary lights.
In the late 19th century, artists such as Georges Seurat, Paul Signac, and Vincent van
Gogh invented a technique of painting that came to be known as pointillism (Figure 6.10).
Brightness: the perceived
intensity of the light present
Lightness: the psychological
experience of the amount of light
that gets reflected by a surface
Additive color mixing: the
creation of a new color by a
process that adds one set of
wavelengths to another set of
wavelengths
Subtractive color mixing:
color mixing in which a new
color is made by the removal
of wavelengths from a light
with a broad spectrum of
wavelengths
FIGURE 6.9 Additive Color Mixing
Predicting what color will be created when mixing monochromatic lights
involves using the color circle. If you are mixing two lights, you just draw a
straight line from one color along the perimeter to the one you are mixing
it with. The color you create will lie somewhere along that line, depending
on the brightness of each light. To predict what color will be created when
mixing three monochromatic lights, you can connect the three colors along
the perimeter to create a triangle as seen in the figure. The color created
will fall somewhere inside that triangle. Where it falls will depend on the
respective degrees of brightness of the three monochromatic lights.
Green Yellow
25% green + 25% blue
Cyan
50% cyan + 50% red
Blue Magenta
Red
80% red + 20% green
45% red + 45% green + 10% blue
No
nsp
ect
ral
pur
ple
s
ISLE 6.4
Color Mixing
FIGURE 6.8 Newton’s Prism Experiment
By Sascha G
rusche (O
w
n w
ork) [CC BY-SA

4.0 (http://creativecom
m
ons.org/licenses/by-
sa/4.0)], via W
ikim
edia Com
m
ons

158 Sensation and Perception
In pointillism, an artist uses small distinct dots of simple pri-
mary colors as the basis of a painting. From a distance, the
dots of colors blend together through a process similar to
additive color mixing to form a rich array of colors. What
makes pointillism relevant here is that it uses additive color
mixing in painting rather than subtractive color mixing, the
approach typically used in painting. The difference is that
colors are created not by mixing paints but by keeping each
dot a specific color, so that the dots blend together in a per-
son’s vision when viewed from a distance.
Subtractive Color Mixing
(Mixing Paints)
Subtractive color mixing is more common in the natural
world, as there are many more situations in which pigments
on the surfaces of objects interact than situations in which
lights interact, as outside of artificial lighting, all lighting
in nature comes from the sun and is therefore white light,
not monochromatic light. Subtractive color mixing occurs
when we mix substances with different absorption spectra.
That is, when we mix substances, the mixture will absorb
the wavelengths both substances absorb, leaving only those
wavelengths to reflect that both do not absorb. For exam-
ple, consider repainting a room in your house. Perhaps as a
young child, you wanted pink walls. Your parents may have
painted your room with a very low saturation pink. Pink, as
a low-saturation red, absorbs most wavelengths but reflects
red light more than it does other wavelengths. As paints are
likely to have a bit of spread in the wavelengths they reflect,
there is also some yellow light that may be reflected (Figure
6.11). Now you want your room to be a light shade of green.
So, when the first coat of paint is applied, you use a light
green paint on top of the existing pink. To your dismay, when
you look at the swath of color on your wall, you have a
drab yellow wall instead of a light green one. How does this
occur? Indeed, the green paint may include a little reflection
of light in the yellow zone as well. So, when the two paints
are mixed, green pigment now absorbs the red wavelengths
that were previously being reflected, and the old pink paint
absorbs the green wavelengths from the new paint. What is
left is the yellow in between. You can see a demonstration of
subtractive color mixing on ISLE 6.4b.
Color-Matching Experiments
Color mixing and matching can be quite a complex phenom-
enon. Nonetheless, understanding how we match colors is
critical to the development of theories of human color vision,
which is the focus of much of this chapter. Therefore, before
we introduce the physiology of color vision, we briefly discuss
color-matching experiments. Our contention here is that with a
M
et
ro
po
lit
an
M
us
eu
m
o
f a
rt
, o
nl
in
e
co
lle
ct
io
n
FIGURE 6.10
A Sunday Afternoon on the Island of La Grande Jatte
This famous painting by Georges Seurat (1859–1891) illustrates the
principles of pointillism. In pointillism, an artist uses small dots of color
to form bigger images. With respect to color, the dots may be of very few
primary colors, but when seen from a distance, the colors form an additive
mixture, and we see a richer array of colors.
FIGURE 6.11 Subtractive Color Mixing
When light green paint is applied to the pink wall, the green paint
continues to absorb most wavelengths other than green and, to some
extent, close-by colors such as yellow. The pink paint continues to
absorb most wavelengths other than red and, to some extent, close-by
colors such as yellow. So when the first coat of green paint is put on the
wall, the net result may be a yellow color.

159 Chapter 6: Color Perception
mix of three primary (monochromatic) col-
ors, we can re-create most any other color.
This is how televisions and computer moni-
tors can reproduce any color: by mixing red,
green, and blue lights. However, there are
other possible primary colors that can be
mixed to re-create other colors. For exam-
ple, color printers, using subtractive color
mixing, employ cyan, magenta, and yellow
to create a large array of colors. Review
ISLE 6.4, particularly noting the different
primaries used in the additive and subtrac-
tive color mixing versions of the illustration.
Many psychophysical experiments
have looked at the perceptual reality of
color matching. These experiments are
known as metameric color-matching
experiments (Silverstein & Merrifield,
1985). A metamer is a psychophysical
color match between two patches of light
that have different sets of wavelengths.
This means that a metamer consists of two
patches of light that look identical to us in
color but are made up of different phys-
ical combinations of wavelengths. For
example, a monochromatic light at 520
nm looks green. But we can create a color
that looks identical by balancing the right
amounts of light at 420, 480, and 620 nm.
In an experiment in metameric match-
ing, an observer is shown two patches
of light. One is called the test patch, and
the other is called the comparison patch.
The test patch is a single wavelength of
an unchanging illumination or brightness
(e.g., a monochromatic light at 520 nm at
a set brightness). The comparison patch is
composed of three primary colors, such
as the red, green, and blue used in a tele-
vision. The observer has control over the
intensities of the three lights in the com-
parison patch. The observer’s task is to
adjust the level of each of the three pri-
mary colors in order to make the color of
the comparison patch equal to that of the
test patch. When the observer does this to
his or her satisfaction, a metameric match
is achieved for that individual. In general, a match that satisfies one individual will also
satisfy other individuals of the most common level of color-vision abilities. That is, most
observers will see this match as identical unless they have some form of color deficiency or
the four-cone system described earlier. This type of matching is illustrated in Figure 6.12.
You can try metameric matching for yourself on ISLE 6.5.
Metamer: a psychophysical
color match between two
patches of light that have
different sets of wavelengths
Test patch
(monochromatic
light)
Test patch Comparison patch
Dials to adjust
color intensities
of three primary lights
Monochromatic
light (test color)
Comparison patch
(mixture of three
monochromatic lights)
Blue
Green
Red
FIGURE 6.12 Metameric Matching
The participant in a metameric matching study is shown a test patch of monochromatic light.
The participant has control over three primary lights in the comparison patch. He must adjust
the intensity of each of the primary lights until the mix of the three lights looks subjectively
identical to the monochromatic test patch.

160 Sensation and Perception
In Figure 6.12, you can see the test patch with a wavelength of 550 nm (light green).
The observer has control over three primary lights, with wavelengths of 440 (dark
blue), 490 (dark green), and 650 (red) nm. These three primary lights are projected onto
the comparison patch and combined through additive color mixing (they are lights, not
pigments or paints). The observer has control over the intensities of the three primary
lights making up the comparison patch and can change them until a metameric match is
made. Even though none of the comparison patch primary lights is the same as the test
patch light, we see the two patches as identical. This is the essence of metameric match-
ing and, to some extent, the basis of much of the color vision we experience, given the
amount of time modern people spend looking at computer screens, cell phone screens,
and television screens.
TEST YOUR KNOWLEDGE
1. Define a metamer, and describe how color-matching experiments contribute to
this concept.
2. Contrast additive and subtractive color mixing.
THE RETINA AND COLOR
6.4 Examine the role of having three cone classes for our perception of color.
In Chapter 3, we introduced the cone systems in the retinae of the eyes (see Table 3.2).
Cones are photoreceptors in the foveae of the retinae that are responsible for high acu-
ity and color vision. Here we explain how cones allow us to code for color and how
they transmit information about color to the occipital lobe. First, we review the func-
tion of the three cone systems and then describe how the presence of three cone types is
important in developing the trichromatic theory of color vision.
Most of us have three classes of cone photoreceptors present in the foveae of the
retinae. Each cone type has a different photopigment present and is therefore sensitive
to a different band of wavelengths. The cone that has a maximum response to light at
420 nm is known as the S-cone (because it is sensitive to short-wavelength light). It
is sometimes (and erroneously) called the blue cone because 420-nm light is perceived
as blue. Calling it the blue cone is misleading because, as we discussed with metameric
matches, all three cones are critical to perceiving all colors. The M-cone class has a
maximum response to light at 535 nm, and the L-cone class has a maximum response
to light at 565 nm. Here the M stands for medium, and the L stands for long, even
though the two peaks are surprisingly close to each other. For the M-cone, 535-nm light
is a yellowish green, and for the L-cone, 565-nm light is still to the yellow side of red.
When these cone systems are combined together, we can see color over a range of 400
to 700 nm, from the S-cone’s sensitivity to shorter wavelengths to the L-cone’s sensi-
tivity to longer wavelengths of light. This range is approximate—there have been some
studies suggesting that some people can detect very bright light at levels just below 400
nm. Nonetheless, the 400- to 700-nm range is a pretty good estimate. This is illustrated
graphically in Figure 6.13.
The S-cones are distinctive for several reasons. First, there are many more M- and
L-cones than there are S-cones. Indeed, S-cones make up only 5% of the total number of
cones. Second, S-cones are less sensitive to light overall than are M- and L-cones. This
means that they are less important in our perception of brightness, but they are still
S-cone: the cone with its peak
sensitivity to short-wavelength
light, around 420 nm (blue)
M-cone: the cone with its
peak sensitivity to medium-
wavelength light, around 535
nm (green)
L-cone: the cone with its peak
sensitivity to long-wavelength
light, around 565 nm (yellow)
ISLE 6.5
Color-Matching Experiment:
Metameric Matches

161 Chapter 6: Color Perception
very important in our perception of color,
especially when we consider their role in
opponent processes. Interestingly, M- and
L-cones are not all identical. There are
subclasses of each (Mollon, 1992). These
differences may result in subtle differences
in color perception from person to person.
To understand how the cone systems
relate to color perception, consider the fol-
lowing. A monochromatic light at 500 nm
(green) is projected onto a white piece of
paper. When the reflected light strikes the
retina, there will be a very weak response
in the S-cone, a strong response in the
M-cone, and a relatively weak response in
the L-cone (Figure 6.14). It is this pattern
of responses that induces the experience
of the color green. In metameric match-
ing, we duplicate the pattern of responses
of the three cone classes with other pri-
mary wavelengths. Thus, we can dim the lowest wavelength light and strengthen the
higher wavelength lights of three primary colors to create the same pattern of responses
of the retinal cones. If the output of the cones is the same (it does not matter if it is
caused by physically different patches of light), the two patches will look the same. You
can try this for yourself on ISLE 6.6.
We have just reviewed the finding that each cone type has a wavelength to which it
maximally responds. As we can see in Figures 6.12 and 6.13, each cone type responds
to a swath of different frequencies as well. The
M-cone, for example, responds weakly to light at
450 nm, greatly to light at 535 nm, and weakly
to light at 650 nm. If this is the case, how can
the M-cone distinguish between a very bright
light at 450 nm and a dimmer light at 535 nm?
The answer is that it cannot. If this were all the
information provided to the V1, a person could
not see in color. Any cone system, by itself, can-
not determine wavelength and therefore color.
Furthermore, consider the response of the M-cone
to a light equally intense at 500 and 630 nm.
Given that the M-cone responds equally strongly
to these wavelengths, equally intense lights at
these frequencies cannot be distinguished by the
M-cone. Thus, at least two cone types are neces-
sary for any color vision to occur.
Univariance, or Why More Than One
Receptor Is Necessary to See in Color
The principle of univariance means that any single cone system is color-blind in the
sense that different combinations of wavelength and intensity can result in the same
FIGURE 6.13 Peak Sensitivity of Cones
The sensitivities of the S-, M-, and L-cones are given as a function of their response
sensitivities to light of different wavelengths. The rod system’s sensitivity is also shown for
comparison. An object’s color is determined by the joint response of each cone in response
to that object’s reflected wavelength pattern.
420
100
50
0
N
o
rm
a
li
ze
d
a
b
so
rb
a
n
ce
400 500
Wavelength (nm)
600 700
535 565
LMS Rod
498
ISLE 6.6
Trichromatic Theory and
Cone Responses
FIGURE 6.14 The Response of Cones to a 500-nm Light
Note that each cone system responds to this light but with a weaker or stronger
response. Color is partially determined by this pattern of responses of each cone
to any particular wavelength.
R
e
la
ti
ve
s
e
n
si
ti
vi
ty
R
e
la
ti
ve
r
e
sp
o
n
se
r
a
te
Wavelength (nm)
Short Middle
Cones
Long300
0
1.0
0
1.0
LongMiddleShort
750
Univariance: the principle
whereby any single cone system
is color-blind, in the sense
that different combinations of
wavelength and intensity can
result in the same response from
the cone system

162 Sensation and Perception
response from the cone system. This implies that color vision depends critically on the
comparative inputs of the different cone systems. Indeed, we can define the problem of
univariance in the following way: The problem of univariance refers to the fact that
many different wavelength and intensity combinations can elicit the same response
from a single cone system. This means that any individual cone system cannot, by itself,
distinguish colors.
The solution to this problem is that we have three cone systems that allow us to
map any particular wavelength at any particular intensity onto a specific pattern among
responses by the three cone systems. A monochromatic light, for example, will elicit a
unique combination of outputs of the three cone systems that distinguishes it from other
monochromatic light. However, we can still “trick” a three-cone system by adjusting
the intensity of two or more lights so that they elicit the same overall response from the
cone systems. This “trick” is the metameric match created by our television monitors.
In summary, a metamer is the matching of a monochromatic color by manipulating the
intensity of three lights so that their combination of wavelengths and intensities results
in the same net output as that of the monochromatic light.
Recall the women whose retinae possess four cones, introduced at the beginning of
the chapter. The reason these women can perceive more colors than the rest of us is that
the fourth cone has a different spectral sensitivity than do the rest of the cones. Thus,
any pattern of light will elicit a more complex response by the four cones. For these
people, the intensity of four lights must be adjusted to match any particular wavelength.
This leads them to have millions more combinations of settings of their four-cone sys-
tems than do people with three-cone systems.
Finally, the problem of univariance explains why we do not see colors under night-
time lighting conditions. When we are under scotopic conditions (dim light), we use
only our rod system. Because we have one class of rods, we do not see color under these
conditions. You can see a dynamic illustration of viewing colored lights under one-cone
or rod-only vision on ISLE 6.7.
TEST YOUR KNOWLEDGE
1. Diagram the differences between the different cones that we have.
2. Why are three cones necessary to see the colors we do? What evidence is
there that some people have a four-cone system?
THE TRICHROMATIC
THEORY OF COLOR VISION
6.5
Describe the trichromatic theory of color vision and its
relationship to the three classes of cones.
Although there have been many theories of color vision throughout history, the first
modern theory was proposed by Thomas Young, the 19th-century philosopher and
scientist, and later further developed by Hermann von Helmholtz, perhaps the 19th
century’s greatest scientist (Figure 6.15). Both of these scientists advanced a theory that
is similar to what we call today the trichromatic theory (and sometimes the Young–
Helmholtz trichromatic theory). Breaking down the word trichromatic, you have tri,
which means “three,” and chromatic, which means “colored.” This theory proposes that
color vision is based on there being three elements in our visual system that respond
ISLE 6.7
Univariance and Color
Matching in Monochromat
or During Scotopic Vision

163 Chapter 6: Color Perception
differently to different wavelengths. Young and
Helmholtz did not know about cones; the theory
was instead developed to explain how we make
color matches. Now, however, it fits in squarely
with the idea of a three-cone retina. The trichro-
matic theory of color vision states that the color
of any light is determined by the output of the
three cone systems in our retinae.
We have already reviewed much of the evidence
that supports the trichromatic theory. First, col-
or-matching experiments show that it takes a mini-
mum of three primary colors to make a metameric
match to a single monochromatic light. For people
with three-cone vision, mixing two lights will not
always yield a match to a pure monochromatic
light, and four is not necessary (for nearly all peo-
ple). That it takes three lights to make a match sup-
ports the trichromatic theory and was observed by
Helmholtz, though he did not yet have a physiolog-
ical explanation. Second, there is strong evidence
for the trichromatic theory from the overwhelm-
ing data supporting the existence of three classes of cones in human retinae. Third, the
trichromatic theory predicts what happens when individuals lose one of the cone classes
in their retinae. These people—sometimes called color-blind but more accurately called
color deficient—still see in color, but they cannot distinguish between hues as three-coned
individuals can. In the classic form, most color-deficient people cannot distinguish between
reds and greens. There are now ample data showing that red–green color deficiency results
from the loss of either the M-cones or the L-cones. There is also evidence that shows that
blue–yellow color deficiency is the result of loss of the S-cones. Furthermore, a color-defi-
cient individual with only two cones requires a match of only two light sources to match
any one monochromatic light. The bottom line is that the trichromatic theory explains
extremely well the relation of cones in the retinae and our perception of color.
TEST YOUR KNOWLEDGE
1. Explain how the fact that we have three cones is related to the three primaries
of the trichromatic color system.
2. If having three cones is necessary for having three primaries, how many
primaries would be needed to match colors in a person with two cones?
THE OPPONENT-PROCESS
THEORY OF COLOR PERCEPTION
6.6
Illustrate the opponent-process theory of color vision focusing on what this
theory can explain about color vision that trichromatic theory cannot.
The traditional rival of trichromatic theory is the opponent-process theory of color per-
ception. Opponent-process theory is based on a different set of observations and has a
different underlying physiological basis. Of course, we now know that the two theories
N
ational Portrait G
allery, London
©
G
L A
rchive/A
lam
y
Trichromatic theory of color
vision: the theory that the
color of any light is determined
by the output of the three cone
systems in our retinae
FIGURE 6.15 Thomas Young (1773–1829) and Hermann von
Helmholtz (1821–1894)
Thomas Young (a) and Hermann von Helmholtz (b) developed the trichromatic
theory of vision. This theory was a precursor of the modern view of color vision, in
which the cones essentially serve as a trichromatic system.
(a) (b)

164 Sensation and Perception
are not mutually exclusive, and that both explain some
aspects of color vision. But until relatively recently,
they were considered rival theories. Certainly, when the
opponent-process theory was first developed, it chal-
lenged the Young–Helmholtz view.
The historical rival of Hermann von Helmholtz was
Ewald Hering, as discussed in Chapter 1. Helmholtz
proposed his version of trichromatic theory, and Hering
proposed the opponent-process theory of color per-
ception (Figure 6.16). We know now that each theory
explains certain aspects of color perception that the
other theory cannot account for. We have already seen
how trichromatic theory explains what the cones of
the retinae are doing. We will see shortly how oppo-
nent-process theory explains color perception, starting
in the retinal ganglion cells and continuing to the lat-
eral geniculate nucleus (LGN) and the occipital cortex.
For many years, opponent-process theory was discred-
ited, but the work of Russell De Valois on opponent
cells in the LGN led to its revival (e.g., De Valois, 1960,
1965; De Valois, Abramov, & Jacobs, 1966). But before
we discuss the physiological evidence for opponent-
process theory, we will discuss the perceptual data.
Findings That Support Opponent-Process Theory
Findings such as color-naming data and complementary color data (discussed later)
led Hering to propose that color vision was not trichromatic but was organized with
four primaries, or unique hues. These four primaries are organized in two sets of
oppositional pairs (blue–yellow and red–green). Blue and yellow are opposite to each
other, and red and green are also opposite to each other. In Hering’s view, there
were several observations that Helmholtz’s trichromatic theory could not explain.
Nonetheless, in their day, Helmholtz’s view prevailed. But we now have physiological
evidence to support opponent-process theory, in addition to Hering’s perceptual data.
Here are some of the important reasons advanced to support opponent processes in
color perception:
a. Nonprimary colors can look like combinations of two primary colors. For
example, purple looks like a combination of blue and red, and orange looks
like a mix of yellow and red. But no colors look like a mix of blue and yellow
or red and green. That is, our perception of colors supports the idea that red
and green do not combine and that blue and yellow do not combine. It is hard
to imagine what a yellowish blue might look like or what a reddish green might
look like.
b. In color-sorting experiments, people tend to sort colors into four basic groups—
green, red, yellow, and blue—rather than the three colors that might be predicted
from trichromatic theory. This finding has been observed across Western and
non-Western cultures (Rosch, 1973).
c. Color afterimages are visual images that are seen after an actual visual stimulus
has been removed. If you stare at a bright incandescent light bulb for even a
Opponent-process theory
of color perception: the
theory that color perception
arises from three opponent
mechanisms, for red–green,
blue–yellow, and black–white
Afterimages: visual images
that are seen after an actual
visual stimulus has been
removed
Complementary colors:
colors on the opposite side
of the color wheel that when
added together in equal
intensity give a white or gray
or black
FIGURE 6.16 Hering’s Model of Opponent Processes
Hering’s view was that all colors on the color circle (here the inner circle)
can be represented by two pairs of opposing colors: blue–yellow and red–
green. These opponents are represented along the outer circle.
Green
Red
BlueYellow
ISLE 6.8
Color Aftereffect

165 Chapter 6: Color Perception
short time and then close your eyes, you will continue to see
an afterimage of that light bulb for a relatively brief period
of time. But you will notice that although the light bulb has
a yellowish hue, your afterimage will appear somewhat blue.
Afterimages are strong after seeing lights, but they work for
reflected light as well. After staring at a patch of red light for
30 seconds or so, one can close one’s eyes and see a vague
patch of cyan (between green and blue, at about 485 nm).
These afterimages have led to the amusing illustrations you
can see in Figures 6.17 and 6.18. An afterimage is seen as
the complementary color, which are colors on the oppo-
site side of the color wheel. When mixed together in equal
intensity, you get a white or gray or black. Complementary
colors are not the same as an opponent color. Red and
green are opponent colors, but because
when added together you see a yellow,
not a white, they are not complements
(review ISLE 6.4a). Thus, aftereffects
lead to complementary colors, whose
existence is nonetheless a problem for
trichromatic theory. For a demonstra-
tion of afterimages, go to ISLE 6.8.
d. Simultaneous color contrast occurs
when our perception of one color is
affected by a color that surrounds it.
The effect occurs when a color is sur-
rounded by its opponent color and not
by other colors or achromatic back-
grounds. A green square will seem
more green if it is surrounded by a red
background, and a red square will seem
more red if it is surrounded by a green
background. Similarly, a blue square
will seem more blue if it is surrounded
by a yellow background, and a yellow
square will appear more yellow if it
is surrounded by a blue background
(Figure 6.19). Notice how the appearance of the central square color changes
toward the complementary color of the surrounding color. You can explore an
interactive illustration of simultaneous color contrast on ISLE 6.9.
In a series of important experiments, Hurvich and Jameson (1957) described the
phenomenon of hue cancellation, which led to a rebirth in the idea that opponent pro-
cessing does explain color vision. We turn to this experiment next.
Hue Cancellation
In this classic experiment, the husband-and-wife team of Hurvich and Jameson rein-
troduced Hering’s view on opponency. Hurvich and Jameson (1957) developed a new
way to empirically examine this view. In this experiment, participants saw a mono-
chromatic light at a wavelength between two particular primary colors. For example,
ISLE 6.9
Simultaneous Color Contrast
Simultaneous color contrast:
a phenomenon that occurs
when our perception of one
color is affected by a color that
surrounds it
Hue cancellation: an
experiment in which observers
cancel out the perception of a
particular color by adding light
of the opponent color
FIGURE 6.17 Afterimage of Vivid Pink
Rhododendron Flower and Blue Sky
FIGURE 6.18 Stare at one flag for roughly 30 seconds without moving your
eyes. Then look at a blank space on the page. You should see the correct flag as an
afterimage. Note that complementary colors of afterimages are not the same as
opponents. To get the red in each flag, you need a cyan, not a green.
FIGURE 6.19 Simultaneous Color Contrast
Opponent colors can enhance the experience of each other. Thus, a green square
surrounded by red looks more green than if surrounded by a neutral color. Similarly,
a yellow square looks more yellow when surrounded by a blue background than a
neutral background.

166 Sensation and Perception
a participant might see a cyan light with a wavelength at 485 nm.
Then the participant was given control over the amount of a second
light that could be added to the first, through additive color mixing.
The instructions were to cancel out the blue so that the light appeared
only as green. The participant could do this by adding a yellow-wave-
length light, but if she tried to do it with red light, she could never
succeed. Only yellow light could cancel out the blue light (and vice
versa). However, if you have a light between red and yellow, green must
be added to cancel out the red. You can see the procedure illustrated
in Figure 6.20. You can do a dynamic illustration of this principle on
ISLE 6.10.
We discuss one last bit of psychophysical data on opponent-process
theory before turning to the underlying physiology. Most colors can
be described in terms of combinations of other colors. Orange, for
example, feels like a mix of yellow and red. Pink feels like a mix of
red and white. However, four colors can be described only in terms
of themselves. These four colors are the four opponent colors: blue,
yellow, red, and green. In this way, these colors can be thought of as
unique colors. Interestingly, these unique colors feel basic to us, but
none of them reflects the peak sensitivity of one of our cones. Even
the S-cone’s peak wavelength sensitivity would not be described as an
iconic blue. Rather, 420 nm appears violet, as if there were a bit of red
to it. The peak sensitivity of the M-cone is a yellowish green, and the
peak sensitivity of the L-cone is an orange-yellow. Thus, these unique
colors must arise from a different level of the nervous system than the
cones in the retinae.
Opponent Cells in the LGN and V1
In Chapter 4, we discussed center-surround cells in V1. These are
cells that respond best to a spot of light in the center surrounded by
a darker circle. There are also cells that respond to darkness in the
center surrounded by an annulus of light. We discussed how such a
system was important for edge detection and discriminating shapes. A
similar system exists for color that also supports the opponent-process
view. Researchers have also found opponent cells in V1 that respond
to a particular color in the center and respond best when that center
color is surrounded by an annulus of its opponent color. We can see the first oppo-
nent cells in the retinal ganglion cells, but we focus here on opponent cells in the
LGN and in V1.
In the LGN, one finds color-sensitive cells called cone-opponent cells. These cells
respond best when they are excited by the input from one cone in the center but inhib-
ited by the input from another cone type in the surround. For example, in the LGN, one
can find cells that are excited by L-cones in the center but inhibited by M-cones in the
surround. There are also cells that are excited by S-cones in the center but inhibited by
both L- and M-cones in the surround. These cone-opponent cells are likely the building
blocks of the color-opponent system (De Valois, 2004).
In V1, we find cells that are specific not to cones but to colors themselves. The
difference is that cones are linked to wavelength responses, whereas the perception of
color is affected by adaptation and the area around them. Review color aftereffects
ISLE 6.10
Hue Cancellation
Unique colors: colors that
can be described only with a
single color term—red, green,
blue, and yellow
Cone-opponent cells:
neurons that are excited by
the input from one cone type
in the center but inhibited by
the input from another cone
type in the surround
FIGURE 6.20 Hue Cancellation
In a hue cancellation experiment, a participant
starts off with a given color (in the figure, a greenish
blue). Then the participant must add another color to
eliminate any experience of one of the colors (e.g.,
the blue). Thus, the participant must add yellow to
eliminate blue (or red to eliminate green). In these
studies, only the opponent-process theory predicts
how participants match the colors.
Start with green
(a)
Add blue to cancel yellow
Still yellowish
Just right
Add more blue
Add even more blue
Too bluish
Start with green
(b)
Add red to cancel green
Still greenish
Just right
Add more red
Add even more red
Too reddish

167 Chapter 6: Color Perception
Color-opponent cells:
neurons that are excited by
one color in the center and
inhibited by another color in
the surround, or neurons that
are inhibited by one color in
the center and excited by
another color in the surround
Double-opponent cells: cells
that have a center, which
is excited by one color and
inhibited by the other; in the
surround, the pattern is reversed
ISLE 6.11
Single- and Double-
Opponent Cells
FIGURE 6.21
Color-Opponent Cells
(a) A single-opponent cell’s receptive
field is shown. It responds best when
one color (red) is seen in the middle
and its opponent (green) surrounds it.
(b) Different types of cone-opponent
cells get inputs from different types
of cones. There is also a pathway
for lightness. It gets inputs from the
M and L wavelength cones. (c) The
center of each double-opponent cell
is excited by one color (green) and
inhibited by the opponent color (red).
In the surround, the cell is excited
by red and inhibited by green. When
we put this cell just along the edge
of a green–red barrier, it will respond
maximally, whereas when it is in an
open green or red field, the excitation
and inhibition will cancel out. Thus,
the cell along the edge will respond
robustly, whereas the cell in the open
field will not. Thus, opponent cells
serve as color-edge detectors.
and simultaneous color contrast if this distinction is not apparent. These neurons are
called color-opponent cells. Color-opponent cells are excited by one color in the center
and inhibited by its opponent color in the surround. A color-opponent cell may also be
inhibited by one color in the center and excited by its opponent color in the surround
(Figure 6.21). Color-opponent cells are red–green and blue–yellow but never, for exam-
ple, red–yellow or green–blue, consistent with Hering’s opponent-process theory. That
is, these cells work in opponent pairs. There will be a cell that responds to red in the
center and inhibits green in the surround, the presumed basis of simultaneous color
contrast. There is also a set of cells that track lightness, the difference along the gray
scale, from black to white. Figure 6.21b shows how the different types of cones connect
to these different color-opponent cells.
In V1, there is also a class of color-sensitive cells called double-opponent cells.
Double-opponent cells have a center, which is excited by one color and inhibited
by the other. In the surround, the pattern is reversed. Thus, if the center is excited by
green and inhibited by red, the surround will be excited by red and inhibited by green.
Double-opponent cells are useful for detecting color edges, that is, where one colored
object ends and a differently colored object begins, by enhancing color divisions at
the edges of objects (Johnson, Hawken, & Shapley, 2001). Think of red berries sur-
rounded by green leaves. Double-opponent cells sharpen the boundaries where the red
berries end and the green leaves begin, making it easier for the observer to home in
on some delicious raspberries. This type of information is useful to the visual system.
For a demonstration of how single-opponent and double-opponent cells work, go to
ISLE 6.11.
TEST YOUR KNOWLEDGE
1. Survey the evidence in support of opponent-process color theory.
2. What are afterimages? Why do they support opponent-process color
theory?
Panel b Source: Adapted from Karen K. DeValois and Michael A. Webster (2011), Scholarpedia, 6(4):3073. http://
www.scholarpedia.org/article/Color_vision
R+
G− S M L S
S vs LM L vs M L+M
M L S M L
Opponent Opponent Non-Opponent
− − −+ + + +
R−/G+ R−/G+
R−/G+
R+/G−R+/G−
R+/G−
Double-
opponent cell
Neutral
Worst
Neutral
Best
R−/G+
R+/G−
R−/G+
R+/G−
(a) (b)
(c)

http://www.scholarpedia.org/article/Color_vision

http://www.scholarpedia.org/article/Color_vision

168 Sensation and Perception
THE DEVELOPMENT OF COLOR PERCEPTION
6.7
Explain what we know about color vision in infancy
and how it changes as we age.
Color Perception in Infancy
Infants are born with all three types of cones already present in each eye (Xiao &
Hendrickson, 2000). But the presence of the proper receptors does not guarantee com-
plete color perception. From the discussion earlier, it is clear that it takes
more than a complete set of cones to see in full color. The cones are only
the start of the process. We briefly discuss here the processes whereby
infants develop the ability to see and distinguish colors. The first question
is how we get a nonverbal infant to report on differences in something as
subjective as color.
One way to study color perception in infancy is to use a method that
relies on habituation. As humans, we do not like repetition, and this applies
to young infants as well. When stimuli are repeated, we start to ignore a
stimulus, but more than that, we have learning mechanisms that lead us
to stop responding to a repeated stimulus. Think of a neighbor having a
car that backfires. The first time it happens, you might jump. If the car
continues to backfire over time, eventually you will stop responding. This
lack of responding to a repeated stimulus is call habituation. Infants are
like adults in this way; they also stop responding to a repeated stimulus.
When a stimulus is changed, both adults and infants will respond again,
a process known as dishabituation. So, if, after you are used to the car
backfiring, your neighbor decides to start up target practice and fires off
a rifle, you will probably jump again.
In the habituation research paradigm, a stimulus is presented repeatedly
to the infant. Usually the first time the stimulus is presented, the infant will
respond by widening her eyes and staring intently. Over time as the same
stimulus is presented, the infant will stop looking at the stimulus. At this
point, the infant has habituated. The critical part comes next. A new stim-
ulus is presented. If the infant can perceive the difference in the stimulus,
she will again look. If the difference in the stimulus is not perceived, the
infant will continue to ignore the stimuli. This can be used to determine if
color—when all other dimensions are held equal—causes dishabituation.
Bornstein, Kessen, and Weiskopf (1976) used this method to examine color per-
ception in 4-month-old infants. The researchers presented a stimulus of a particular
wavelength, say 510 nm, to the infants repeatedly, 15 times to be exact. A trichromat
would see this as green. Over time, the infants stopped responding. Then, immediately
after these trials, the infants were then presented more trials. The crucial trials are those
that had stimuli with wavelengths that were different from the original. In the condi-
tion with the 510-nm light, the new wavelengths were 480 nm and 540 nm. The two
wavelengths differed the same number of nanometers of wavelength from the 510-nm
stimulus, but perceptually that is not the case. A 540-nm light still appears green (but
with a bit of yellow in it), but the 480-nm light appears blue to adults. To restate this,
510-nm and 540-nm wavelength lights appear similar and might be called green or
greenish, but a 510-nm wavelength light and a 480-nm wavelength light appear very
different: One appears green (510 nm) and one appears blue (480 nm).
Fr
ed
B
ur
re
ll/
Sc
ie
nc
e
So
ur
ce
FIGURE 6.22
Infants Have All Three Types of Cones
Infants are born with all three types of cones. The
question is whether having these cones gives
newborns color vision.
Habituation: the learning
process in which animals
stop responding to a repeated
stimulus
Dishabituation: the process
in which, after habituation
has occurred, changing the
stimulus causes the organism
to respond again

169 Chapter 6: Color Perception
The question then becomes whether infants see these colors in the same way that
adults do. If they do, then the infants should show much stronger dishabituation to the
480-nm stimulus than to the 540-nm stimulus, and that is exactly what Bornstein and
colleagues found. In their study, they also tested the color boundaries between green to
yellow and yellow to red, and infants gave responses that suggest that they have similar
color distinctions to adults. The interesting nature of these findings, combined with
findings from other studies (Brown & Tellar, 1989), suggests that infants have more
than just operating cones; they also have opponent processes for color vision operating.
It should not be taken that infants have completely adultlike color vision at birth. In
fact, the retinae and foveae are quite immature, but the basics of color vision are pres-
ent as early as we can study, and the way their color vision operates is similar to how
adults’ color vision operates.
Aging and Color Perception
So, what happens to color vision as we age? Adults do not lose cones with normal
aging, so the basic mechanisms of color vision seem intact, but that does not mean that
color vision does not change as we age. One way that color vision changes in aging,
even in healthy aging eyes, is through changes to the transparency of the lens. As we
age, light does not pass as easily through the lens. It is not a great change, and if the
loss of transparency were the same at all wavelengths, then the change would be of
little interest. However, the lens loses transparency to short wavelengths more than
for other wavelengths, lessening our ability to see blue colors. The change is not great
in healthy eyes, but there is an increase in confusion of blues and yellows as would be
expected from opponent-process theory (Pinckers, 1980). The loss of the sensitivity to
blue is slow and usually unnoticed as we age. However, if a person develops the need for
an artificial lens, the change can be noticed because, usually, only one eye gets the lens
implant at a time. So, for a period of time, a person will have the artificial lens in one
eye, which will transmit the short wavelengths clearly, and their aging lens in the other
eye. By opening and closing each eye, the differences in seeing color can be very clear.
Granville (1990) had the procedure done and both described his experience and tested
his eye on various color-matching tests. The differences were dramatic in the blue color
range, greatly enhancing his enjoyment of a clear blue sky and even the color of a gas
jet. Granville’s experience was greater than for those who do not require implants, but
the type of changes would follow the same pattern.
TEST YOUR KNOWLEDGE
1. Describe how the process of habituation is used to determine if infants have
color vision.
2. Diagram how the loss of transparency to short wavelengths causes yellow–blue
confusions in older adults.
COLOR DEFICIENCY
6.8
Diagram the different types of color deficiency and explain why this
term is, almost always, a better term than color blindness.
Color deficiency refers to the condition of individuals who are missing one or more of
their cone systems. Color deficiency is often called color blindness, but the latter term
Color deficiency: the
condition of individuals who
are missing one or more of
their cone systems

170 Sensation and Perception
is rarely appropriate. The vast majority of color-deficient individuals do see colors,
just not as many colors as trichromatic people do. Color deficiencies are usually the
result of genetic variations that prevent the development of one or more cone systems.
Thus, in general, color deficiencies are present at birth and seldom develop later in life.
Interestingly, because of the nature of those genetic issues, color deficiencies are much
more common in men than they are in women.
The genetic information for forming cones in fetal development travels on the
X chromosome in a region of that chromosome disproportionately likely to have a
variation (Woods & Krantz, 2001). These genetic variations are of different types, and
it is the type of genetic information present that determines the type of color deficiency.
It is important to remember that males have only one X chromosome and females have
two. Because the genes that lead to color deficiency are on the X chromosome, males
need only one of these color-deficient genes to be color deficient, but females need two
of these genes, one on each chromosome, to be color deficient. As a result, males are far
more likely than females to have a color deficiency. Such types of traits are known as
sex-linked traits. Other sex-linked traits include male pattern baldness and hemophilia,
both much more common among men than women.
Color deficiency is quite common in humans. Usually, in a college class of 100 stu-
dents or so, there is at least 1 young man who admits to being color deficient. Indeed, it
is estimated to occur in as many as 8% of males and 0.4% of females (Birch, 2001). The
most common diagnosed form is red–green color deficiency. These individuals see colors
in a general spread from dark blues to light blues to dim yellows to bright yellows. The
frequencies three-cone people see as green and red are just part of the blue-to-yellow con-
tinuum for individuals with this form of color deficiency. However, there are really two
separable types of red–green color deficiency, one caused by the loss of the M-cone system
and one caused by the loss of the L-cone system. In the past, a simple way to identify color
deficiency was to show potential patients Ishihara plates
(Figure 6.23). In these plates, trichromatic people can see
the numbers created by the dots because they are different
colors than the surrounding dots. The dots, however, are
isoluminant dots; that is, they reflect light at the same
intensity, even though the wavelengths are different for
the dots that make up the number and the background.
Because of this, trichromatic people will see the number,
because they can detect the color differences. But a red–
green color-deficient individual will see only dots and not
be able to tell them apart because of intensity differences.
Because the dots are isoluminant and chosen to be meta-
mers for color-deficient individuals, a red–green color-de-
ficient person cannot see the number in an Ishihara plate.
Thus, these plates allow for easy identification of color
deficiency. Another way of identifying color-deficient
individuals is to have them do metameric matching tasks.
Their matches will be very different from those of people
with three-cone vision. You can try some different tests
for color deficiency in ISLE 6.12. It is important to note
that because these tests are run under a monitor and not
the conditions called for in the test, their results are only
approximate.
To best be able to understand the different types
of color deficiencies, we start by reminding you about
trichromatic color vision and then how the different
ISLE 6.12
Color Deficiency Tests
©
P
RI
SM
A
A
RC
H
IV
O
/A
la
m
y
FIGURE 6.23 An Ishihara Plate
These figures are used as a quick test for color deficiencies. The dots are
all isoluminant, regardless of color, so the number cannot be determined
by differences in brightness. Only color allows you to see the number. For
the plate shown, a red–green color-deficient individual would not be able
to see the number.

171 Chapter 6: Color Perception
color deficiencies differ from it. People with trichromatic color vision have three func-
tioning cone systems. Such a trichromatic person could match any pure wavelength by
varying the intensity of three colors, such as red, green, and blue. Across trichromatic
people, these matches will be nearly identical. Then there are color-deficient people,
who come in a number of varieties.
Rod Monochromacy
Rod monochromacy is a very rare form of color deficiency, affecting only 1 in every
30,000 people (Carroll, Neitz, Hofer, Neitz, & Williams, 2004). Rod monochromats
have no functioning cones of any kind and therefore can be described as truly color-
blind. As a result, they see the world in shades of gray—high-reflectance objects are
white, low-reflectance objects are black, and intermediate-reflectance objects are vari-
ous shades of gray. In metamer matching, only one color is required to match another,
as all a rod monochromat will be doing is adjusting the percentage of reflected light,
that is, how gray the surface is.
Because they have no cones, rod monochromats have many other visual problems
in addition to color blindness. Rod monochromats are dependent on their rod vision in
both bright and dim light. This has serious disadvantages during daylight conditions, as
these individuals are highly sensitive to light but have poor visual acuity. Because they
are using scotopic vision all the time, rod monochromats are extremely sensitive to bright
lights. For example, in a room that would be considered normally illuminated for people
with intact cone systems, a rod monochromat will find it too bright. As such, rod mono-
chromats often must wear sunglasses indoors. Going outside on a bright sunny day can
be overwhelming and requires very strong eye protection. In reality, most rod monochro-
mats will avoid bright outdoor conditions even with very strong sunglasses. Moreover,
because the rods do not support acuity, rod monochromats have very poor visual acuity
and must wear very strong lenses in order to read. Even then, most rod monochromats
require large typefaces in order to read normally. To summarize, unlike those with color
deficiencies, rod monochromats are at a serious disadvantage relative to people who have
cones as receptors as well as rods (you can see the world as a rod monochromat might on
ISLE 6.13). You can read about the experience of people with this disorder in a book, The
Island of the Colorblind, written by famed neurologist Oliver Sacks (1998), who visited
a Pacific Ocean island where this disorder is relatively common.
Cone Monochromacy
There are extremely rare cases of individuals known as cone monochromats. These
people lack two cone types but have one present. S-cone monochromats have been
found; they have the S-cone system but lack the M- and L-cone systems (Alpern, Lee,
Maaseidvaag, & Miller, 1971). S-cone monochromacy is an X-chromosome-linked trait
and thus is more common in men than women, though it has been observed in women.
However, S-cone monochromacy is still extremely rare in men, as only 1 in 100,000
men will exhibit it. Because of the low overall sensitivity of the S-cone system, cone
monochromats exhibit many of the symptoms seen in rod monochromats, although
the symptoms tend to be less severe in S-cone monochromats than in rod monochro-
mats. Cone monochromats also have poor acuity and high sensitivity to bright light.
Wavelength discrimination is poor in S-cone monochromats, and subjectively, the world
appears in blacks, whites, and grays. One of the authors (JHK) took his sensation and
perception class years ago with a fellow student who had this type of color deficiency.
Despite taking the class in the evening and night (6 pm to 9 pm once a week), the fellow
student wore sunglasses and the lights were kept as low in the room as the class could
ISLE 6.13
Rod Monochromat Vision

172 Sensation and Perception
tolerate. As night came on, the student became more comfort-
able and sometimes took her sunglasses off. Interestingly, in
twilight conditions, in which both the rod and the cone sys-
tems are at work, S-cone monochromats can distinguish some
colors, essentially using the one cone system and the one rod
system as might a dichromat.
Dichromacy
Dichromats have only two working cone systems. Thus, they
can see colors, though a much lesser range of colors than
do trichromats. There are three major forms of dichromacy
depending on which cone is absent: protanopia, deuterano-
pia, and tritanopia (Table 6.1). Dichromats require only two
colored lights to match any monochromatic light, compared
with trichromatic individuals, who require three. Dichromats
see in color, but they cannot make some of the discriminations
that are easy for trichromats. Protanopia and deuteranopia
are linked to the X chromosome and are therefore inheritable
and more common in men than women. Protanopia and deuteranopia are also more
common than tritanopia.
Because men have only one X chromosome (and one Y chromosome), if they have
a deficiency gene on the X chromosome, they will express the deficiency. However,
women have two X chromosomes. The same deficiency gene must be present on both X
chromosomes for a woman to be a protanope or a deuteranope. If a woman has only
one color-deficient X chromosome, she will see colors as a trichromat, as the single
trichromatic X chromosome is enough to structure her retinae correctly. However, she
is still a carrier, and thus there is a 50% chance that she will pass that gene on to an
offspring. A son inheriting the carrier X chromosome would be color deficient, and a
daughter would have a 50% chance of being a carrier. Thus, the only way a woman can
be color deficient is if she has a color-deficient father and either a color-deficient mother
or a mother who is a carrier.
Protanopia
Protanopes lack L-cones in their retinae, as a function of a deficient gene. Protanopia
can occur in as many as 1% of males but is very rare in females, roughly 0.02%.
Because protanopes do not have L-cones, they are classified as red–green color-blind
(the common name for the condition). An approximation of what they see as a function
of wavelength is depicted in Figure 6.24a. At short wavelengths, protanopes see blue.
As wavelength increases, the blue becomes less saturated until it eventually becomes
gray (at 492 nm), and then as the wavelengths continue to increase, the color is per-
ceived as a more and more saturated yellow. The yellow fades at the high end of the
visual spectrum (Figure 6.25). This 492-nm light is called the crossover wavelength for
protanopes.
Deuteranopia
Deuteranopes lack M-cones in their retinae, as a function of a deficient gene.
Deuteranopia has about the same frequency as protanopia, and in many contexts, they
are indistinguishable and are both referred to as red–green color blindness. However,
Protanopia: a lack of
L-cones, leading to red–green
deficiency; this trait is sex-
linked and thus more common
in men
Deuteranopia: a lack of
M-cones, leading to red–
green deficiency; this trait
is sex-linked and thus more
common in men
Tritanopia: a lack of S-cones,
leading to blue–yellow color
deficiency; this trait is rare
and is not sex-linked
Anomalous trichromacy: a
condition in which all three
cone systems are intact, but
one or more has an altered
absorption pattern, leading to
different metameric matches
than in the most common type
of trichromatic individuals
TABLE 6.1
A Description of Different Types of Color Vision
Type of Color
Deficiency
Cone System
Absent
Color
Experience
Rod
monochromacy
All No color
Tritanopia S Lack green and
yellow
Deuteranopia M Lack green and
red
Protanopia L Lack green and
red
Anomalous
trichromacy
All present, but
one is different
from standard
Near standard
color vision
Common
trichromacy
None Standard color
experience

173 Chapter 6: Color Perception
FIGURE 6.24 The World as It Looks to Color-Deficient Individuals
This is how these colored displays look to (a) protanopes, (b) deuteranopes, and (c) tritanopes.
FIGURE 6.25 The Visual Spectrum as It Appears to Color-
Deficient Individuals
400
492
700
400 700
400 700
400 500 600 700
Protanope
Deuteranope
Tritanope
498
570
the crossover wavelength from blue to yellow occurs at a different wavelength (498
nm). For an approximation of how deuteranopes perceive color, inspect Figure 6.24b.
Tritanopia
Tritanopes lack S-cones in their retinae. This is the
rarest form of dichromat color deficiency, occurring
in just about 1 in 100,000 people. Because this defi-
ciency is not sex-linked, it is just about as common
among women as it is among men. Tritanopes see
blue at short wavelengths, which becomes less sat-
urated as the wavelength increases. At higher wave-
lengths, tritanopes see red. The crossover point for
tritanopes is 570 nm. For an approximation of how
tritanopes perceive color, inspect Figure 6.24c. You
can explore a set of interactive illustrations of dichro-
macy on ISLE 6.14. You can examine the cones
missing and how that impacts color matching in dif-
ferent types of dichromacy, and you can simulate the
appearance of images for dichromats. You can try
your own photographs in the ISLE demonstration.
There is also a form of color deficiency called
anomalous trichromacy. In anomalous trichro-
mats, all three cone systems are intact, but one or
more of the cone systems has an altered absorp-
tion pattern, leading to different metameric
matches than in the most common form of cones
in trichromatic individuals. The most common
form of anomalous trichromacy is an abnormality
in the M-cone system (Smith & Pokorny, 2003).
However, it can also occur with the L- or the
S-cone system. In general, anomalous trichromats
ISLE 6.14
Dichromacy
(a) (b) (c)

174 Sensation and Perception
are not as good at discriminating similar wavelengths as the more com-
mon type of trichromats. You can see the different forms of anomalous
trichromacy and their incidence in Table 6.2.
A question that may have occurred to you is how we can be sure of
the colors dichromats and anomalous trichromats are seeing. After all,
because they have missing or abnormal cone systems, their perception of
color may not be imaginable by people with trichromatic color vision. If
this were the case, the perceptual spectra in Figures 6.24 and 6.25 would
not be accurate. In other words, when looking at the sky, a deuteranope
may know to say “blue” as trichromatic individuals do, but is the col-
or-deficient person really seeing the same color?
The answer to this question comes from a small class of color-
deficient persons known as unilateral dichromats. A person with unilateral dichromacy
has dichromacy in one eye but trichromatic vision in the other. This is an extremely rare
condition. However, the existence of unilateral dichromacy allows researchers to discover
the nature of the experience of dichromacy. A unilateral dichromat can observe a colored
figure with one eye at a time and describe the color both as a person with trichromatic
color vision would and as a person with dichromacy would (Alpern, Kitahara, & Krantz,
1983; Graham & Hsia, 1958). When viewing objects with both eyes, they see color essen-
tially as a trichromat would, but when they close their trichromatic eye, they become
dichromats. Thus, such a person would describe a 520-nm light as green with her trichro-
matic eye but as yellow with her dichromatic eye. When she reopens her trichromatic eye,
she again sees it as a fully trichromatic person would. It is based on the experiences of
these unilateral dichromats that we know the color experience of dichromats.
As we discussed at the beginning of the chapter, some women seemingly have four-
cone systems. Interestingly, these women are usually related to men with anomalous
trichromacy, so one theory that accounts for the existence of this condition is that the
same mutation that results in anomalous trichromacy in men can occasionally result
in tetrachromacy in women (Jordan et al., 2010). According to this research, roughly
10% of women may carry the gene for anomalous trichromacy. But in Jordan et al.’s
(2010) study, only 1 in 24 women with this gene showed any evidence of being a tetra-
chromat. Such women require four separate colors to match a monochromatic light and
presumably see countless shades of color that trichromatic individuals do not (Jameson
et al., 2001). To tetrachromats, all of us trichromatic individuals are effectively color
deficient. Tetrachromacy is extraordinarily rare, but it has been documented.
Cortical Achromatopsia
In cortical achromatopsia, loss of color vision occurs because of damage to the occip-
ital lobe. Cortical achromatopsia is extremely rare and is much less common than the
color deficiencies caused by cone problems in the retinae. Cortical achromatopsia usually
comes about from damage to area V4 (also involved in shape perception, as discussed in
Chapter 5). Cortical achromatopsia usually involves a perceptual experience of seeing
only in black and white (and shades of gray) or the impression that colors seem washed
out. In some cases, the perception is of shades of gray, but patients can still discriminate
by wavelength even though they do not experience the colors. These patients still have
all their cone systems intact, and the problem is at a higher level of processing. This may
be why they can discriminate wavelengths but still not see colors. In addition, in some
cases, patients with achromatopsia lose the ability to remember color. That is, they do
not remember what it was like to experience color. In contrast, a person who becomes
blind because of eye damage still remembers his or her experience of color. In some cases,
patients with cortical achromatopsia can no longer put color into mental images and may
fail to be able to remember objects by their colors. Thus, a banana is no longer yellow in
TABLE 6.2 Percentages of Different
Anomalous Color Deficiencies
Type of
Anomalous
Trichromacy
Males Females
Protanomalous
(L-different)
1% 0.02%
Deuteranomalous
(M-different)
5% 0.04%
Tritanomalous
(S-different)
0.005% 0.01%
Unilateral dichromacy: the
presence of dichromacy in
one eye but trichromatic vision
in the other
Cortical achromatopsia: loss
of color vision due to damage
to the occipital lobe

175 Chapter 6: Color Perception
memory, nor is a ripe tomato red. Even in memory, these objects lose their colors. Finally,
in some cases, individuals may not even be aware that color vision has been lost (von Arx,
Müri, Heinemann, Hess, & Nyffeler, 2010). Again, all of this symptomatology has been
linked to damage in V4 (Wade, Augath, Logothetis, & Wandell, 2008).
TEST YOUR KNOWLEDGE
1. Diagram the different types of color deficiencies.
2. What is the difference between a rod monochromat and a cone monochromat?
What differences are there in their vision?
CONSTANCY: LIGHTNESS
AND COLOR CONSTANCY
6.9 Assess the term constancy and how it applies to color vision.
Constancy refers to our ability to perceive an object as the same
object under different lighting and environmental conditions. For
example, we want our visual system to recognize a barracuda
whether it is swimming directly toward us or away from us, in
murky water or clear water, at an angle or coming straight at us.
The image on the retinae may be quite different in each of these
cases, but we still need to recognize the same object. This is a vital
goal of the visual system. We want to be able to tell if an object is
the same object across changes in lighting, shading, distance, move-
ment, and orientation. For another example, we want to be able to
detect a banana as being a banana whether we are seeing it close up
or from far away. We want to be able to judge if the banana is ripe
regardless of whether we are viewing the fruit in broad daylight, by
the light of a light bulb, or even by twilight. For this reason, we have
evolved many processes that enact constancy in different domains
of perception. Lightness constancy refers to our ability to perceive
the relative reflectance of objects despite changes in illumination.
Color constancy refers to our ability to perceive the color of an
object despite changes in the amount and nature of illumination.
Before we delve into the science of color constancy and then
lightness constancy in depth, consider Figure 6.26. We see a pleasant
garden scene. Look at the grass. We see a light green field of grass
with a few scattered small white flowers. In the background, some pretty yellow flowers
adorn a bush. Across the foreground, a tree, which is not visible in the photograph, casts a
shadow over the grass. The core idea of constancy is that we see the grass as green, and the
same green, despite the changes in illumination caused by the shade. We perceive correctly
that there is a shadow present and infer that there is differential illumination on the grass, but
the grass’s color does not change as a function of illumination. This is an illustration of color
constancy, the fact that we see an object as the same color across changes in illumination.
Color Constancy
We consider color constancy first. Color constancy refers to the observation that we
see the same color despite changes in the composition of the wavelengths of light that
Constancy: the ability
to perceive an object as
the same under different
conditions
Lightness constancy: the
ability to perceive the relative
reflectance of objects despite
changes in illumination
Color constancy: the ability
to perceive the color of an
object despite changes in
the amount and nature of
illumination
FIGURE 6.26 Color Constancy
We want to be able to recognize the grass here as one
continuous object of one continuous color, even though some
of the grass is shaded by trees and therefore reflects less light
back to the viewer. That we see the grass as one color despite
the differences of illumination is an example of constancy.

176 Sensation and Perception
is striking that object. Thus, a green mug appears to be the same color regardless of
whether the light illuminating it is natural sunlight, a fluorescent light bulb, or an
incandescent light bulb. This is true even though the object is now reflecting different
absolute amounts of light at different wavelengths under each illumination condition.
Color constancy serves an important perceptual function. The properties of objects
seldom change as a function of changes in the source of illumination. Thus, a system
that sees an object as a constant color across such changes leads to accurate perception.
Interestingly, the distribution of wavelengths in sunlight changes across the day. Evening
light has more long-wavelength light than light earlier in the day. Although we might
enjoy the colors of twilight, we do not normally see objects as changing colors, though
we are aware of general changes in illumination when we attend to them. We can see
this in Figure 6.27. The statues of the presidents appear to be the same color despite
the change from peak sunlight to twilight. You can explore an interactive illustration of
these principles on ISLE 6.15.
Color constancy is not perfect. There are a number of situations in which color con-
stancy fails; that is, we see the same object as being differently colored under different
lighting conditions. One such situation is when we use a monochromatic light. Shining
a monochromatic light on an object and not the area around the object will allow that
object to reflect only that wavelength. Thus, depending on how much of that wave-
length is reflected by the object, the object will appear the color of the light regardless
of its reflectance characteristics. What we will see is a different level of saturation,
depending on how much the object reflects the wavelength being shined on it. Another
situation occurs when an object is viewed in front of a pure black background. This
makes it difficult for the visual system to get the context to see the object as the right
color. Simultaneous color contrast illustrates another way we can fool color constancy
(see Figure 6.19 and ISLE 6.9).
The mechanism whereby our visual systems control color constancy is not well
understood. Seemingly, the visual system is able to compare the reflectance patterns
of different wavelengths from one object with those from another to determine which
object is reflecting more blue, yellow, and so on. However, because the visual system
cannot intrinsically measure the wavelength distributions in the illuminant light, it must
infer this as well (Foster, 2011). Data support the idea that the visual system automati-
cally determines the ratio of wavelengths in a scene by comparing across many objects
in that scene. That is, the color of an object is determined not just by the wavelengths
©
iS
to
ck
ph
ot
o.
co
m
/S
un
Ch
an
©
iS
to
ck
ph
ot
o.
co
m
/j
ov
an
ni
g
ISLE 6.15
Illumination and Color Constancy
FIGURE 6.27 Color Constancy on Mount Rushmore
We see the monument as being the same color despite the change in illumination from day to night.
(a) (b)

177 Chapter 6: Color Perception
coming off of that object but also by the wavelengths of the light coming
from neighboring objects.
One of the classic demonstrations of color constancy has been called
Mondrian experiments by Edwin Land (1977), after the Dutch painter
Piet Mondrian (Figure 6.28). In these experiments, a surface with a ran-
dom collection of rectangles of different sizes and colors of matte paper
(much like Mondrian’s paintings) is presented to participants. Each rect-
angle is surrounded by rectangles of different colors. The illumination
of the surface is provided by three projectors, each of which projects a
different narrow band of wavelengths. One projector emits long wave-
lengths, the second emits middle wavelengths, and the third emits short
wavelengths. When all give light with the same level of illumination,
these three projectors are a match for white light.
The method for these experiments is complicated. First, a partici-
pant is asked to look at one square, say a red one. Once the participant
is focused on the red square, the experimenter changes the illumination
by adjusting the three projectors so that the light coming from the red
square reflects twice as much middle-wavelength light as long-wave-
length light. The question for the participant is what is the color now
for this red square? The square itself reflects more long-wavelength
light than others (hence its red color). But now, it is being differentially
illuminated by more light shorter in wavelength. After the adjustment
of the projectors, the reverse is true; it is reflecting more middle-wave-
length light than long-wavelength light. Still, participants say that the
rectangle they have been focused on the whole time looks red. Within
wide ranges of the bands of illumination, the wavelength coming from
the square does not change its color appearance. Once red, always
red. And of course, the same is true for squares of other colors on
the Mondrian image as well. This is important because the other
squares have changed their reflectance as a function of the incoming
lights. The fact that each square is surrounded by squares of differ-
ent colors is important to the outcome of this experiment. Thus, the
comparison of each area with the varied colors around it is critical to
maintaining color constancy. This supports the idea that color constancy is achieved
by an implicit comparison across different objects in a scene, each object reflecting
a different wavelength, allowing the visual system to extract the illumination from
known reflective properties.
Lightness Constancy
Lightness constancy refers to our ability to perceive the relative reflectance of objects
despite changes in illumination. The easiest way to think of lightness constancy is to
think of it along the continuum from black to gray to white. These achromatic colors
simply refer to the amount of white light an object reflects. A black object absorbs most
light, whereas a white object reflects most light, with gray objects being in between.
Lightness constancy refers to the observation that we continue to see an object in terms
of the proportion of light it reflects rather than the total amount of light it reflects. That
is, a gray object will be seen as gray across wide changes in illumination. A white object
remains white in a dim room, while a black object remains black in a well-lit room.
In this sense, lightness constancy serves a similar function as color constancy in that it
allows us to see properties of objects as being the same under different conditions of
lighting (Adelson, 1993).
©
iStockphoto.com
/hohl
FIGURE 6.28 Building Painted in the
Style of Piet Mondrian (1872–1944)
This building painted in the style of Mondrian shows a
pattern of color. Stimuli such as this have been used for
experiments on color constancy. In the experiments,
the wavelengths of the light illuminating the stimuli are
altered. As long as viewers can see all the different
boxes of color, they maintain color constancy. But if
they only see one square, its color will change as a
function of illumination.

178 Sensation and Perception
Consider an object that reflects 25% of the light that hits its surface.
This object will be seen as a rather dark gray (Figure 6.29). If we leave it in
a dim room that receives only 100 units of light, it will reflect 25 units of
light. However, if we place it a room that is better lit, it will still reflect the
same 25%. If there are now 1,000 units of light, it will reflect 250 units of
light. But we still see it as the same gray, despite the fact that the object is
reflecting much more light. Similarly, an object that reflects 75% of ambi-
ent light will be seen as a light gray in the dim room, even though it reflects
less total light than it does in the bright room. Thus, lightness constancy
is the principle that we respond to the proportion of light reflected by an
object rather than the total light reflected by an object. You can explore an
interactive illustration of these principles on ISLE 6.16.
As with all constancies, there are conditions that create illusions that
overwhelm the processes that create constancy. The Gelb effect is an
exception to the principle of lightness constancy. The Gelb effect is a
phenomenon whereby an intensely lit black object appears to be gray
or white in a homogeneously dark space. Think of a cat caught in the
headlights of a car at night. The headlights illuminate only a small space
in front of the car, dark pavement in addition to the cat. Because there is
nothing to compare the object to, the cat appears white, because it is reflecting a lot of light
in an otherwise dark space. However, if we suddenly place a white object next to the cat,
the cat now appears black, its actual color, as our visual systems now have something with
which to compare it. For an illustration of the Gelb effect and related illusions, go to ISLE
6.17. The Gelb effect neatly shows the importance of the ratio principle in explaining light-
ness constancy (Gilchrist et al., 1999).
The ratio principle states that the perceived lightness of an object is explained by
the ratio of light it reflects rather than the absolute amount of light it reflects, assuming
even illumination across the visual scene. As long as illumination is constant across
the field of view, the ratio captures the properties of reflectance. In an illusion such
as the Gelb effect, the constant illumination requirement is violated—the cat is more
illuminated than its surroundings, and lightness constancy is compromised. Of course,
situations such as the Gelb effect are rare in the real world, and for the most part, the
ratio principle allows us to correctly interpret the lightness of objects.
TEST YOUR KNOWLEDGE
1. Why is it important to have color constancy in the real world? What function
does it serve?
2. What is the difference between color constancy and lightness constancy?
ISLE 6.16
Lightness Constancy
ISLE 6.17
Gelb Effect
Gelb effect: a phenomenon
whereby an intensely lit black
object appears to be gray or
white in a homogeneously
dark space
©
E
dw
ar
d
H
. A
de
ls
on
FIGURE 6.29 Lightness Constancy
In this illustration, we see a checkerboard being shaded
by a peculiar cylinder. Because we infer that there is less
illumination on the shaded part of the checkerboard, we
see square A as being the same shade as square B, even
though objectively B is a darker gray than A.
EXPLORATION: The Color Purple
Purple has, in the West, long been associated with royalty
and religiosity. Perhaps because of its rarity in nature and
the difficulty in generating purple dyes, purple has always
been a regal and expensive color. We find this from our
earliest recorded records to recent history. In the Hebrew
Bible, God tells Moses that the Israelites should bring
Him offerings of purple cloth. In both ancient Greece and
Rome, purple was the color worn by kings and royalty, a
tradition that continued in Europe through the centuries.
Roman Catholic bishops wear purple to this day. This
regard for purple is also seen in other cultures—purple
is a symbol of royalty in Japan and of piety in China.

179 Chapter 6: Color Perception
Even today, purple dyes made from murex snails can cost
hundreds of dollars. Because purple pigments are also rare
in nature, we value plants and flowers that are purple as
well. Purple flowers are a staple of gardens everywhere
(Figure 6.30).
Following Livingstone (2002), we make a distinction
between violet and purple. In practice, these colors are
very similar perceptually, but technically, they refer to
different kinds of color. Violet is a spectral color that we
see at the very shortest wavelengths of the visual spec-
trum, shorter even than blue. We see violet from 400 to
about 440 nm, at which point, wavelengths become blue.
Purple, by contrast, is a nonspectral color that cannot be
generated by a single wavelength of light. Purple is a mix
of red light and blue light. Most dyes and pigments are
purple, reflecting both red and blue light, rather than vio-
let (Figure 6.31). We now describe why these colors look
similar (see Monnier, 2008).
At 440 nm and under (i.e., until about 400 nm, below
which we cannot see), colors look blue but also a little
red. That is why these colors are called violet. Although
they are spectral colors, they look like a mix of red and
blue. There is a clear physiological explanation for this
phenomenon. At 440 nm, our S-cones are nearly at their
peak. However, the L-cone system is also active. Although
these cones are sensitive mainly to long-wavelength light,
they have this minor peak for short-wavelength light.
Thus, these very short wavelengths activate both blue
and red responses in our cones, resulting in the percep-
tion of violet.
Purple is a nonspectral color that results from the combi-
nation of red and blue light. A more bluish purple may be
metameric to violet, but purple is a mixed color, not a spec-
tral color. Thus, purple objects need to have an odd pattern
of reflectance. To be purple, an object must absorb light in
the central or green part of the visual spectrum but reflect
light in the red and blue portions at the ends of the visual
spectrum. This particular arrangement—absorbing wave-
lengths in the middle of the spectrum but not at the end-
points—is particularly difficult for chemical compounds to
achieve. Hence, the rarity of purple in nature. You can see
that for purple plants, the surface reflects the long and short
wavelengths but absorbs the middle ones. Because purple is
rare in nature, it also makes it more difficult for people to
generate purple dyes, which, in the past, led to high prices
for purple dyes and paints. Long live purple!
©
iStockphoto.com
/O
G
photo
FIGURE 6.30 Purple Flowers
Purple flowers are especially valued by gardeners everywhere.
FIGURE 6.31 How to Make Purple
To be purple, an object must absorb light in the central or green part of the visual spectrum but reflect light in the red and blue portions at the ends of
the visual spectrum.
Source: From Livingstone, 2002, p. 31.
700650600550
Wavelength (nm)
African violet
African violet
Purple cabbage
R
e
la
ti
v
e
r
e

e
ct
a
n
ce
500450400
0
0.5
1
700650600550
Wavelength (nm)
Ultramarine Blue
Cobalt violet
Ultramarine
violetR
e
la
ti
v
e
r
e

e
ct
a
n
ce
500450400
0
0.5
1

180 Sensation and Perception
APPLICATION: Color Deficiency
and Our Colorful Information
A major technological innovation that was first proposed
in 1931 has changed our world, even if it does not get
the attention that other innovations have. In that year,
an international organization, the Commission inter-
nationale de l’éclairage (International Commission on
Illumination), thankfully, CIE for short, proposed a set
of equations to allow for automatic color reproduction
(Silverstein & Merrifield, 1985). In some ways, color
reproduction has been going on since ancient times, but
before this innovation, to reproduce a color in nature
required a human eye, an artist. With this set of equa-
tions, it was possible to start specifying primary values,
like the red, green, and blue elements on a television
screen, to make a metamer for another color. It no longer
required a person but could be done automatically by
a machine. It was not long after this time that movies
started being produced in color, for example, Gone With
the Wind, Fantasia, and The Wizard of Oz, which came
out by the end of the 1930s, and these films depended on
these equations. To color a movie before this innovation,
artists would have to hand paint each and every frame
of a movie, which did occur, but very rarely. With time,
we got color printing, color television, and all the other
ways that color is part of our lives now. But this advance
has come at a cost. The CIE equations work for trichro-
mats but not dichromats or other people with color
deficiencies. Let us take a look as some of the difficulties
raised for dichromats.
Conventionally, stoplights all around the world use red
for stop and green for go. How do deuteranopes and pro-
tanopes, who are “red–green color-blind,” know to stop
at a traffic signal or drive through, if they cannot discrim-
inate between red and green? This is more of a practical
question than one related to the nature of their visual
experience. In the past, traffic signals were a significant
problem for color-deficient drivers. However, they are no
longer a major problem because it is now standard to
put the red light at the top of the series of lights (three if
there is a yellow light) and the green light at the bottom.
In addition, they use a red and green that will appear
different lightnesses to dichromats as well (Proctor &
Van Zandt, 2008). So, color-deficient individuals can stop
and go by attending to vertical position rather than color.
Those of us with trichromacy do not pay attention to this
feature, but color-deficient people learn it quickly and
use it to respond to traffic signals. Although this posi-
tioning helps, many dichromats still report some trouble
with traffic lights, and some authors argue that automo-
bile drivers should meet some level of color vision ability
before being allowed to drive (Cole, 2016). This argu-
ment would add driving to those jobs that require a cer-
tain level and type of color vision present to be able to
be hired, such as pilots and railroad engineers. In both
of these cases, although trichromatic color vision is not
required, at least anymore, the applicant must be able to
discriminate between certain colors easily. For example,
railroad engineers must be able to discriminate red, green,
and yellow—a requirement that rules out most prota-
nopes and deuteranopes.
The news is not all bad for dichromats, though. In fact,
it is surprising there are so many of them given all of the
difficulties we have outlined in this chapter. Perhaps they
have some sort of advantage that helps them evolutionarily.
Morgan, Adam, and Mollon (1992) discovered that dichro-
mats can see through some forms of color camouflage
that completely mask objects to color trichromats. There
had been anecdotes since World War II about color-defi-
cient people seeing objects hidden by camouflage, but this
study was the first to demonstrate this ability in the lab.
Participants had to locate a square of horizontal rectan-
gles in a field of vertical rectangles. If the dots are all one
Ro
be
rt
B
ro
ok
/S
ci
en
ce
S
ou
rc
e
FIGURE 6.32
Color Deficiency and Our Colorful Information
View through a car windscreen by someone with color blindness, in
particular, deuteranopia or protanopia.

181 Chapter 6: Color Perception
color (Figure 6.33a), the square is easy to find for all of us.
When the figure is camouflaged with a random array of red
and green dots (Figure 6.33b), trichromats perform prac-
tically at chance, whereas dichromats still perform quite
well. You can try a version of this experiment in ISLE 6.18.
You might be asking why this works. Remember the CIE
equations. It works for trichromats and is used to develop
the camouflage. If your color vision is not trichromatic,
the metamers that are being developed for the camouflage
do not work for you. You might recall from your biology
classes that camouflage coloring occurs not just in human
warfare but also in nature as well. It could be the ability
to see through this natural camouflage is an advantage
for dichromats. No, not an
advantage that makes up for
all of the disadvantage, but
enough of one for them to sur-
vive in reasonable numbers.
ISLE 6.18
Color Camouflage
and Dichromacy
FIGURE 6.33
Color Camouflage
(a) It is easy in this version of the
image to find the small box of
horizontal rectangles. (b) With the
addition of a color camouflage,
finding the square becomes much
harder. Try it when the stimulus is
only briefly presented. However,
dichromats have much less trouble
seeing through this camouflage.
CHAPTER SUMMARY
6.1
Examine the relation of the wavelength of light to
perceived color.
The visible spectrum is the band of wavelengths from
400 to 700 nm that people with trichromatic vision can
detect. We see different wavelengths as different colors.
However, wavelength is not the only property that governs
color vision.
6.2
Diagram the perceptual principles of hue, satu-
ration, and brightness and the role they play in
our perception of color.
Hue refers to the color quality of the light and corre-
sponds to the color names we use. Saturation refers
to the purity of the light or the amount of white light
mixed with the colored light. And brightness refers to
the amount of light present. Additive color mixing is the
creation of a new color by a process that adds one set
of wavelengths to another set of wavelengths. When
we add all of the different wavelengths of sunlight,
we see white light rather than many individual colors.
Subtractive color mixing is the creation of a new color
by the removal of wavelengths from a light with a broad
spectrum of wavelengths.
6.3
Formulate the idea of a metamer in additive and
subtractive color mixing in terms of what we
learn from color-matching experiments.
A metamer is a psychophysical color match between
two patches of light that have different sets of wave-
lengths. To create a metamer, there are two types of
color mixing, additive and subtractive. In additive color
mixing, wavelengths from two or more sources of light
are placed close to make the new color. In subtrac-
tive color mixing, wavelengths from a light source are
removed to yield the color match. From color-matching
experiments, we learn that any three wavelengths can
be adjusted to match a single-wavelength light in people
with trichromatic color vision.
(a) (b)

Sensation and Perception182
6.4
Examine the role of our having three cone classes
for our perception of color.
Color vision starts off with the cones of the retina. We
have three cone systems, responsible for different parts
of the visual spectrum. The S-cones have peak sensitiv-
ity to short-wavelength light, at around 420 nm (blue). The
M-cones have peak sensitivity to medium-wavelength
light, at around 535 nm (green), and the L-cones have
peak sensitivity to long-wavelength light, at around 565
nm (yellow).
6.5
Describe the trichromatic theory of color vision
and its relationship to the three classes of cones.
The existence of these three cone types supports the
trichromatic theory of color vision, which states that the
color of any light is determined by the output of the three
cone systems in our retinae.
6.6
Illustrate the opponent-process theory of color
vision focusing on what this theory can explain
about color vision that trichromatic theory cannot.
There is also evidence that supports the opponent-
process theory of color vision, which states that color
perception arises from three opponent mechanisms, for
red–green, blue–yellow, and black–white. Four primaries
are organized in two sets of oppositional pairs (blue–
yellow and red–green). Blue and yellow are opposite to
each other, and red and green are also opposite to each
other. Evidence for this theory comes from color after-
images and hue cancellation studies. Neuroscience has
also shown that there are cone-opponent cells in the
LGN and color-opponent cells in V1. In particular, double-
opponent cells seem to be specialized for detecting
edges, where one color ends and another color starts.
6.7
Explain what we know about color vision in
infancy and how it changes as we age.
Infant vision can be tested using the process of habitu-
ation, which is where we stop responding to repeated
stimuli. Bornstein and colleagues used this procedure
to determine that infants seem to have color categories
similar to the ones that adults do. In other words, infants
dishabituate, that is, respond again, when a color bound-
ary is crossed, say from blue to green, and the boundaries
are similar to those found for adults. In aging, one way
color vision changes is that the lens loses its transpar-
ency to short wavelengths more than other wavelengths,
which causes us to be less able to see blue than other
colors. The change is slow and not great in healthy eyes.
6.8
Diagram the different types of color deficiency
and explain why this term is, almost always, a
better term than color blindness.
Color deficiency refers to the condition of individuals
who are missing one or more of their cone systems. Rod
monochromats have no functioning cones of any kind
and therefore can be described as truly color-blind. Cone
monochromats are people lacking two cone types but have
one present in addition to their rod systems. Protanopia
(red–green color deficiency) is a lack of L-cones, a con-
dition that is sex-linked and therefore more common in
men. Deuteranopia (red–green color deficiency) is a lack
of M-cones, a condition that is sex-linked and more com-
mon in men. Tritanopia (blue–yellow color deficiency) is
a lack of S-cones; this condition is much rarer than the
others and is not sex-linked. Anomalous trichromats
have all three cone systems intact, but one or more has
an altered absorption pattern, leading to different meta-
meric matches than in trichromatic individuals. A unilat-
eral dichromat is a person with dichromacy in one eye but
trichromatic vision in the other. Cortical achromatopsia is
a loss of color vision that occurs because of damage to
the occipital lobe.
6.9
Assess the term constancy and how it applies to
color vision.
Constancy refers to our ability to perceive an object as
the same object under different conditions. Lightness
constancy refers to our ability to perceive the relative
reflectance of objects despite changes in illumination.
Color constancy refers to our ability to perceive the color
of an object despite changes in the amount and nature of
illumination. The Gelb effect or Gelb illusion is an excep-
tion to lightness constancy in which we see as white a
dark object when it alone is illuminated and we cannot
see other objects.
REVIEW QUESTIONS
1. What is the range in nanometers of the human visi-
ble spectrum? What is the difference between het-
erochromatic light and monochromatic light?
2. What is meant by the terms hue, saturation, and bright-
ness? What does each contribute to our perception of

Chapter 6: Color Perception 183
color? Give an example of two colors that differ with
respect to saturation.
3. What are additive color mixing and subtractive color
mixing? How do the two processes differ? When
would you use each one?
4. What are the three cone systems? What kind of mono-
chromatic light is each cone sensitive to? Why do the cone
systems support the trichromatic view of color vision?
5. What is the problem of univariance? How does it
relate to metameric matching? How does it relate to
color blindness under scotopic conditions?
6. What is the opponent-process theory of color vision?
Describe three perceptual phenomena that demon-
strate the reality of opponent processing.
7. What is the difference between cone-opponent cells
and color-opponent cells? Where do you find each
kind of cell? Describe the visual field of a double-
opponent cell.
8. What is meant by the term color deficiency? What
is the difference between a rod monochromat and
a cone monochromat? What are the three types
of dichromats? Describe the physiological issue
and the perceptual consequences for each kind of
dichromat.
9. What is meant by the term constancy? What are
lightness constancy and color constancy? How do
we infer the nature of the illuminant in each case?
10. What is the difference between violet and purple?
What is the physiological explanation for why purple
looks like a mix of red and blue?
PONDER FURTHER
1. Color vision adds a great richness to our visual life.
It also adds a great deal of complexity to our visual
system as you can tell by this chapter. So, what advan-
tages, other than enriching our experience, does
having color vision add to our survival? Recall from
earlier, that most of our fellow mammals have limited
or no color vision.
2. Use ISLE 6.14c and see if you can describe how dif-
ferent types of dichromats will experience the world.
What scenes would a protanope find the most differ-
ent from a trichromat versus a dichromat?
KEY TERMS
Additive color mixing, 156
Afterimages, 164
Anomalous trichromacy, 173
Brightness, 156
Color constancy, 175
Color deficiency, 169
Color-opponent cells, 167
Complementary colors, 165
Cone-opponent cells, 166
Constancy, 175
Cortical achromatopsia, 174
Deuteranopia, 172
Dishabituation, 168
Double-opponent cells, 167
Gelb effect, 178
Habituation, 168
Heterochromatic light, 153
Hue, 155
Hue cancellation, 165
L-cone, 160
Lightness, 156
Lightness constancy, 175
M-cone, 160
Metamer, 159
Monochromatic light, 153
Opponent-process theory
of color perception, 164
Protanopia, 172
Quality, 155
Saturation, 155
S-cone, 160
Simultaneous color contrast, 165
Spectral reflectance, 153
Subtractive color mixing, 156
Trichromatic theory of color
vision, 163
Tritanopia, 172
Unilateral dichromacy, 174
Unique colors, 166
Univariance, 161
Visual spectrum (visible
spectrum), 153

Sensation and Perception184
Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e
SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review,
study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning.
Learning Objectives Digital Resources
6.1 Examine the relation of the wavelength of light to perceived color. How We See Color
6.2 Diagram the perceptual principles of hue, saturation, and
brightness and the role they play in our perception of color.
Focal Colors and Unique Hues
6.3 Formulate the idea of a metamer in additive and subtractive
color mixing in terms of what we learn from color-matching
experiments.
6.4 Examine the role of having three cone classes for our
perception of color.
Color Vision, Cones, and Color-Coding in the Cortex
6.5 Describe the trichromatic theory of color vision and its
relationship to the three classes of cones.
Cone Contributions to Color Constancy
6.6 Illustrate the opponent-process theory of color vision focusing
on what this theory can explain about color vision that
trichromatic theory cannot.
Synesthesia for Color Is Linked to Improved Color
Perception but Reduced Motion Perception
6.7 Explain what we know about color vision in infancy and how it
changes as we age.
6.8 Diagram the different types of color deficiency and explain why
this term is, almost always, a better term than color blindness.
For One Artist, Colorblindness Opened Up a World of Black
and White
6.9 Assess the term constancy and how it applies to color vision. Lightness Constancy Illusion
These X’s Are the Same Shade, so What Does That Say
About Color?
Checker Shadow Illusion by Edward H. Adelson
Optical Illusions Show How We See—Beau Lotto

David Parker/Science Source

7Depth and Size Perception
Tim Flach/Iconica/Getty Images
LEARNING OBJECTIVES
7.1
Explain monocular depth cues and how they work to
allow us to perceive depth with one eye.
7.2 Compare the different oculomotor curves in their effectiveness as depth cues.
7.3
Summarize the principle of stereopsis and how it
applies to human depth perception.
7.4 Sketch the correspondence problem and how it must be solved for stereopsis.
7.5
Explain the concept of size perception and the
inferential nature of its determination.
7.6 Diagram the concept of size constancy and how it functions in our visual system.
7.7
Demonstrate your understanding of illusions of size and depth and
how they help us understand the operation of our visual system.
INTRODUCTION
Our ability to see depth and distance is a skill we take for granted, as is the case for
much of perception. Like other perceptual processes, it is an extremely complex process
both computationally and physiologically. The problem our visual system starts off
with is simple. How do you extract information about a three-dimensional (3D) world
from the flat two-dimensional (2D) surface known as the retina (Figure 7.1)? As we
will see, our visual system uses a complex combination of various cues or clues to infer
depth, which often results in vivid perceptual experiences, such as when the monster
lunges out at us from the movie screen while we’re watching that notoriously bad 3D
movie The Mummy (2017). As most readers know, an important part of depth percep-
tion comes from the fact that we have two eyes that see the world from slightly different
angles. We know that people blind in one eye may not see depth as well as others, but
they still have depth perception and can judge distances quite well. In the course of this
chapter, we discuss how single-eyed people perceive depth and explore the reasons why
binocular (two-eyed) vision enhances our depth perception.
Our visual system processes visual information, including a comparison between
what the two eyes see, to determine depth, but that is not the only way to see in depth.
Other animals’ visual systems achieve depth perception in very different ways. What
would it be like to see the world through a completely different visual system? We
ISLE EXERCISES
7.1 Depth Perception
and Adanson’s Jumping
Spider
7.2 Monocular Depth
Cues
7.3 Motion Depth Cues
7.4 Accommodation
7.5 Vergence
7.6 Stereopsis
7.7 Binocular Disparity
7.8 The Construction
of Visual Depth With
Binocular Disparity
7.9 Anaglyph
Stereograms
7.10 Random-Dot
Stereograms
7.11 Retinal Image Size
and Distance
7.12 Müller–Lyer Illusion
7.13 Ponzo Illusion
7.14 Ames Room
7.15 Moon Illusion
7.16 Virtual Reality
Photographs
7.17 Virtual Reality and
Therapy

188 Sensation and Perception
cannot necessarily experience what this would
be like, but we can study the mechanisms
involved and compare them with our own
visual system. Consider spiders, a class of ani-
mals that have eight eyes in addition to their
famous eight legs. Do eight eyes give spiders
a more complex system to see depth, just as
some of the four-coned animals we spoke of in
Chapter 6 see more colors? Interestingly, there
have been a few studies on how this class of
spiders see depth, and the system is incredibly
different from our own. We’ll consider one of
these studies examining depth perception in
spiders. Nagata et al. (2012) examined the use of depth information in jumping accu-
racy in a species of spiders known as Adanson’s jumping spiders (Hasarius adansoni).
These spiders have eight eyes; the four frontal eyes are used for hunting (Figure 7.2).
As their name suggests, Adanson’s jumping spiders catch their prey by jumping onto
small insects and eating them. To be effective at this hunting strategy, their jumps must
be precise, which requires good depth perception. If they land in front of or behind their
insect prey, the prey is likely to escape. So, good depth perception is necessary if the
spider is going to eat and survive. Nagata et al. (2012) were interested in what neural
mechanisms the spiders use for depth perception, given the large number of eyes but the
small brain in these spiders. So that they could focus on the two largest eyes, known as
the principal eyes, they rendered the spiders’ other eyes temporarily blind by dabbing
them with black paint. They set out insects for the spiders to catch and then assessed
the accuracy of the spiders’ jumps.
Unlike a human eye, which has only one layer of photosensitive cells on the retina,
the retinae in the principal eyes of these jumping spiders have four distinct photo-
sensitive layers. Each layer is sensitive to a different range of wavelengths, much as
our cones are sensitive to different ranges of wavelengths. When an image is in focus
on one layer of the spider’s retina, it is out of focus on the other layers. Although
you might think this would make their vision blurry, the spiders use the extent to
which the second image is out of focus to determine
the distance they are from objects. Nagata et al. (2012)
showed that the spider’s eyes focus a sharp image on
the first layer of the retina, leaving blurred images on
the subsequent layers. The spider then compares the
sharpness of the first image to the blurriness of subse-
quent images to compute an estimate of depth. Thus,
these spiders use information from different layers of
each retina to compute depth. This fascinating way
of determining depth is quite different from how our
visual system determines depth. (See ISLE 7.1 for a
demonstration of how spiders use this system to deter-
mine depth.) In our visual system, we use the compari-
son of the two images’ positions across the two eyes to
determine depth, so it is a very different way of using
the two eyes to see in depth.
Returning to our own depth perception, look around
your current environment. You probably have your book
(or e-book) about 12 to 20 inches away from your eyes.
FIGURE 7.1 Our Visual Systems Must Re-Create the 3D World
Using an Essentially Flat, 2D Retina
This figure shows the equivalence of size of objects at different distances from the
retina. Everything along this cone projects the same-size image onto a flat retina. The
visual system must then determine relative size and relative distance (depth).
FIGURE 7.2 Adanson’s Jumping Spider
This spider has eight eyes but uses the multiple layers in its retinae to
determine depth. In this photograph, its four frontal eyes, used for hunting,
are easily visible.
©
iS
to
ck
ph
ot
o.
co
m
/e
lth
ar
20
07
ISLE 7.1
Depth Perception and
Adanson’s Jumping Spider

189 Chapter 7: Depth and Size Perception
Beyond that, you may see a table, a window, and clouds drifting in the sky outside
your room. It seems almost trivial that the book is closer to you than the window, and
the window is closer to you than the clouds. All of these observations serve to hide a
rather interesting feature of depth perception. The retina is essentially flat. Although
it is curved along the back of the eye, it is all one layer and can be flattened out very
easily. There is only one layer of receptors (unlike in those pesky spiders). Regardless of
the distance an object is from the eye, light is imaged by the same receptors, so at this
point in the visual system, there is no information about depth. In other words, which
receptors serve to transduce the light into action potentials does not depend on how far
an object is from the eye. Thus, other sources of information about distance must be
used for 3D perception. This information must be found primarily in the visual scene
itself. In addition, for the first time, the role of having two eyes will become apparent, in
terms of both the advantages and the complications that having two eyes cause for us.
The solution that our visual system evolved is to rely on a system of cues or clues for
depth perception. This view of how humans see depth is sometimes known as the cue
approach to depth perception. This approach focuses on the observation that because
information on the retina is 2D, we must infer the third dimension from other cues our
visual system provides. As we discuss shortly, these cues include monocular cues such
as occlusion and relative size, oculomotor cues such as convergence and accommoda-
tion, and, most famously, binocular cues from comparing images from each retina. In
this chapter, we consider the three major factors in depth perception: monocular cues
(including motion cues), oculomotor cues, and binocular cues. We start this survey with
the monocular cues.
It is important to make explicit what we mean by a cue or a clue. In this discus-
sion, we essentially use the terms cue and clue interchangeably. A cue is a factor that
aids you in making a nonconscious and automatic decision. Cues tell us about which
objects are closer and which are farther away. Thus, the cue approach to depth per-
ception emphasizes that we combine information we can use to infer depth given that
we cannot compute it directly. Let us go through different types of cues that we use to
see depth.
MONOCULAR DEPTH CUES
7.1 Explain monocular depth cues and how they work to allow us to perceive
depth with one eye.
Monocular depth cues are the information in the retinal image that gives us informa-
tion about depth and distance but can be inferred from just a single retina (or eye). In
everyday life, of course, people with two functioning eyes can also perceive these cues
with both eyes, but they are just as usable with only one functioning eye. That is, these
are cues that tell us about depth even if we are looking at the world with only one eye.
Try it—close one eye. You can still use vision to distinguish between objects near and
far. Some people describe the world as seeming a bit flatter when using only one eye
than when using two, but we still judge distances accurately. Monocular cues include
pictorial cues, those cues from which we can judge depth from static or nonmoving pic-
tures like a photograph, and movement-based cues, in which moving objects or our
own motion allow us to make inferences about depth and distance, as in a traditional
non-3D movie (Figure 7.3).
We start with the pictorial cues.
Cue approach to depth
perception: the system
whereby depth perception
results from three sources of
information, monocular cues
to depth present in the image,
binocular cues from the
comparison of images in each
eye, and cues from focusing
the eyes, such as vergence
and accommodation
Monocular depth cues:
depth cues that require only
one eye
Pictorial cues: information
about depth that can be
inferred from a static picture
Movement-based cues: cues
about depth that can be seen
with a single eye in which the
inference of distance comes
from motion

190 Sensation and Perception
Occlusion (or Interposition)
Occlusion occurs when one object partially
hides or obstructs the view of a second object.
We infer that the hidden object is farther
away from us than the object that obstructs
it. Consider the whitewater kayakers in Figure
7.4a. In the photograph, the blue helmet of
one kayaker partially occludes the view of the
other kayaker’s boat. From this, we know that
the blue-helmeted kayaker must be in front
of the black-helmeted kayaker. Similarly, the
blue-helmeted kayaker’s paddle occludes the
view of his own lifejacket. From this, we infer
that the paddle is in front of the lifejacket. Such
a scene as this makes the cue of occlusion look
obvious. However, it is still an inference, based
on knowledge we bring to the act of viewing
scenes as can be seen by examining Figure
7.4b and 7.4c, where perceived occlusion is
used to create a powerful illusion on an impossible box. Occlusion provides infor-
mation about only relative position, not absolute distance. In the kayaking photo-
graph, we cannot determine with much accuracy how far apart the two kayakers are.
You can explore an interactive illustration of occlusion and all other monocular cues
on ISLE 7.2.
FIGURE 7.3
A Graphical Depiction of the Different Types of Depth Cues
Each class of depth cues is labeled in a box, with the specific depth cues listed below it.
• Binocular
Disparity
• Accommodation
• Vergence
• Occlusion
• Relative Height
• Relative Size
• Familiar Size
• Linear Perspective
• Texture Gradient
• Atmospheric
Perspective
• Shadowing
• Motion Parallax
• Deletion and
Accretion
• Optic Flow
Monocular BinocularOculomotor
Pictorial Motion
Depth cues
FIGURE 7.4 Occlusion
(a) We know that the kayaker in the yellow boat is in front of the kayaker
in the red boat because the head of the closer kayaker occludes the boat
of the kayaker farther away. What other cues for depth are present in this
photograph? (b) How is this box possible? Some of the edges seem to be
in front but also cross edges that are in the back. (c) The impossible box
revealed. The edges do not cross behind each other. From this angle it is clear
that there are gaps in the front edges making it seem that they go behind the
other edges. This illusion shows how we infer depth from occlusion.
(b) (a)
(c)
ISLE 7.2
Monocular Depth Cues

191 Chapter 7: Depth and Size Perception
Relative Height
Relative height means that objects closer to the horizon are seen as more distant. In
a picture, this means that objects below the horizon are seen as nearer to the viewer
if they are closer to the bottom of the picture, but objects above the horizon are seen
as nearer if they are closer to the top of the picture. To understand relative height,
think of the horizon dividing the world into two roughly equal portions, the ground
below and the sky above. Think about paintings you may have seen. Usually, the
horizon is somewhere near the middle in Western perspective painting. The distant
part of the sky is painted near the middle of the picture, near where the sky and
ground meet at the horizon. The part of the ground and sky near the viewpoint of
the painter, then, occurs at the extremes of the painting, the close part of the ground
near the bottom and the close part of the sky near the top of the painting. Examine
the photograph in Figure 7.5a. The horizon here is where the ocean meets the sky
and is very clear in this photograph. This places the conch shell and the coral rocks
in front of the ocean, relatively close to the viewer (or camera). The rocks closest to
the bottom of the photograph are closest to the viewer. However, look at the clouds
above the horizon. The clouds nearest to the top of the photograph are viewed as
closest to the viewer.
We see the same pattern in the painting depicted in Figure 7.5b. The water in the
river is close to us and is depicted as such by being at the bottom of the painting. At
the top of the painting are clouds looming over us. The horizon is in the center of the
painting, where the blue sky and white clouds meet the distant green hills and more
distant buildings.
Relative Size
Relative size refers to the fact that the more distant an object is, the smaller its image
will be on the retina. Therefore, if there are two identical objects, the one that is far-
ther away will be the one that has a smaller image on the retina. For example, if we
assume that the two kayakers in Figure 7.4a are approximately the same size (the one
in the front is actually slightly shorter than his companion), then the one who has a
Occlusion: a visual cue that
occurs when one object
partially hides or obstructs the
view of a second object; we
infer that the hidden object is
farther away from us than the
object that obstructs it
Relative height: a visual cue
in which objects closer to
the horizon are seen as more
distant
Relative size: the fact that the
more distant an object is, the
smaller the image will be on
the retina
©
iStockphoto.com
/D
enKuvaiev
FIGURE 7.5 Relative Height
(a) The horizon here is where the ocean meets the sky and is very clear in this photograph. This places the conch shell and the coral rocks along
the bottom of the image as relatively close to the viewer. However, above the horizon, objects closer to the horizon are judged as farther away, and
objects toward the top of the image are judged to be closer. (b) An illustration of how an artist uses this rule to create depth in a painting.
(a) (b)

192 Sensation and Perception
smaller image on the retina must be farther away from the viewer. As the image of the
blue-helmeted kayaker is larger in the picture than that of his companion, we assume
that he is closer to the viewer. Take your fingers out and measure the actual size of
the two kayakers. Note that the kayaker in the background does not look abnormal
in any way. Despite creating a smaller image on the retina, the person does not look
oddly small. The normal size of the more distant person is due to a mechanism called
size constancy, which we discuss later, after more of the depth cues have been covered.
Because we infer that this person is farther away, we do not mistake his smaller size on
the retina for his being a smaller person. You can explore an interactive illustration of
relative size on ISLE 7.2a.
Familiar Size
Related to relative size is the cue of familiar size. Familiar size comes into play when
we judge distance on the basis of our existing knowledge of the sizes of objects. Thus,
if we know that a particular object is smaller than another object, but it is taking up
more space on our retina, we assume that the smaller object must be nearer to us and
that the larger object is farther away. Thus, in Figure 7.6a, each object occupies the
same amount of space on the retinae, but the watermelon is judged to be farthest away
because we know that this is the largest fruit. Similarly, in the photograph in Figure 7.6b,
familiar size allows us to perceived that the building is much bigger than the planter
even though the planter takes up more room on the photograph.
FIGURE 7.6 Familiar Size
(a) Familiarity can be used as a cue for distance of depth. In this image, there are
three objects that are familiar to most of us, all of which take up the same space on
the retinae. But because we know that strawberries are smaller than oranges and
oranges are smaller than watermelons, we unconsciously infer that the strawberry is
closest and the watermelon is farthest. (b) Here is a real-world example. The height
of the building is smaller in the photograph than some of the flowers. However, our
experience with buildings and flowers helps us see that the building is larger.
Retinal image
(a)
(b)
Familiar size: the cue
whereby knowing the retinal
size of a familiar object at a
familiar distance allows us to
use that retinal size to infer
distance

193 Chapter 7: Depth and Size Perception
The cue of familiar size is often eclipsed by other cues. For example, in Figure 7.7a,
the presence of children climbing on the lobster informs us that the lobster in this fig-
ure is no ordinary lobster. In Figure 7.7b, we find a gift shop inside a lobster trap, and
finally, in Figure 7.7c, the building in the background informs us that this is no ordinary
boot. We might well wonder whether people in Maine and eastern Canada don’t have
better things to do.
Linear Perspective
Linear perspective is the pictorial depth cue that arises
from the fact that parallel lines appear to converge as
they recede into the distance. From the point of view of
a human observer, parallel lines seem to get closer and
closer to each other as they get farther away. Of course,
you may remember from your high school geometry
class that parallel lines never meet. But perceptually
they do, at the edge of the horizon. To see what this
cue looks like, examine the photograph in Figure 7.8a.
The railroad tracks are parallel, and in this scene, they
go straight through the desert landscape. The linear
perspective cue is that the parallel lines of the railroad
tracks seem farther apart when close to the viewer,
as they take up more space in the image and on our
retinae. They get smaller and closer together higher in
this image, and we extract from this information that
the tracks go off into the distance. This linear perspec-
tive serves as a cue to depth. The larger the distance is
between parallel lines (the tracks), the closer those lines must be. Of course, in Figure
7.8a, there are other monocular depth cues in addition to linear perspective. Figure 7.8b
shows another good example of linear perspective. The numbers all look the same size
to us, even though the lower numbers, seen toward the top of the photograph, are much
FIGURE 7.7 When Familiar Size Is Unhelpful
(a) The presence of children informs us that this is an abnormally large lobster. It is possible that the lobster is normal size, and the children are the
size of shrimp, but this is a less likely scenario, so the presence of the children immediately informs us of the nature of the lobster. (b) Note the normal
lobster traps in front of the building. Various cues here tell us that the restaurant is normal size and the lobster trap that surrounds it is unusually large.
(c) The famous L.L. Bean boot, with various amounts of context that tell us about the boot’s size.
©
D
av
id
R
. F
ra
zi
er
P
ho
to
lib
ra
ry
, I
nc
./
A
la
m
y
©
V
is
io
ns
o
f A
m
er
ic
a,
L
LC
/A
la
m
y (c) (b) (a)
Linear perspective: the
pictorial depth cue that arises
from the fact that parallel lines
appear to converge as they
recede into the distance
FIGURE 7.8 Linear Perspective
In these images, parallel lines appear to converge in the distance.
(a) We infer that the train tracks are parallel and thus must be getting
more distant as they converge toward the top of the photo. (b) In the
photo of the finish line, how many depth cues can you see?
©
D
ig
ita
l V
is
io
n/
Ph
ot
od
is
c/
Th
in
ks
to
ck
©
iS
to
ck
ph
ot
o.
co
m
/m
ur
en
gs
to
ck
ph
ot
o
(a) (b)

194 Sensation and Perception
smaller in terms of their retinal image size, but because
they are farther away, they do not look smaller.
For painters who wish to convey a 3D scene on a 2D
canvas, linear perspective is an important technique. Look
at the painting in Figure 7.9, which is called Paris Street:
A Rainy Day, by Gustave Caillebotte (1848–1894).
Look at the building in the background. The separation
between the floors is clearly indicated. The walls appear
to diverge and give the indication that the building is
angled and larger in width toward the back, and that the
surfaces recede in depth. In fact, the lines that make up
this building are essentially parallel. Thus, parallel lines
indicate a flat surface, and converging lines that we see
as parallel indicate a surface that recedes in depth. Note
also that we assume that the people are roughly the same
size. Thus, the smaller images of people serve as a cue that
those people are farther from the point of the viewer.
Texture Gradients
A texture gradient is a monocular depth cue that occurs
because textures become finer as they recede in the dis-
tance. Texture gradients as a monocular cue are clearly
related to relative size. In both cases, we use existing
knowledge about sizes or patterns of objects and assume
that smaller images closer to the horizon are the same size,
but farther away. Most surfaces, such as walls and roads
and a field of flowers in bloom, have texture. As the sur-
face gets farther away from us, this texture gets finer and
appears smoother (Gibson, 1950). Another way of say-
ing this is that common elements that are evenly spaced
in an image appear more close together in the distance
than they do in the foreground. For example, the tulips
in Figure 7.10 represent a texture gradient. We assume
that the flowers are about the same size and the same
distance apart in the field. That the images are smaller
and closer together toward the top of the image suggests
that these tulips are farther away. We can also see texture
gradients in the cobblestones in the painting depicted in
Figure 7.9. The cobblestones get progressively smaller as
the road recedes in depth, until the stones are not clearly
distinguishable from one another. In the distance, only the
general roughness of the street is noticeable.
Atmospheric Perspective
Atmospheric perspective is a pictorial depth cue that arises from the fact that objects
in the distance appear blurred and tinged with blue. When we look at a visual image,
close objects are clear and well defined, and objects farther away are more blurred.
Moreover, because the atmosphere scatters light, more distant objects will also have a
blue tinge. This feature of depth perception can be seen in Figure 7.11. In the photograph
in Figure 7.11a, the road in the foreground is clear and sharp. The distant mountains at
FIGURE 7.9 Linear Perspective in Art
This painting by Gustave Caillebotte (1848–1894), titled Paris Street:
A Rainy Day, was painted in 1877. Notice how Caillebotte used linear
perspective to show depth and distance in the painting. There is a
progression of people painted large and up close toward the bottom
of the painting and smaller and farther toward the top of the canvas to
indicate their distance from the painter. Notice also the odd building in
the distance. The lines that make up the walls appear to diverge and give
the indication that the building is angled and larger toward the back,
and that the surfaces recede in depth. In fact, the lines that make up this
building are essentially parallel.
G
oo
gl
e
Cu
ltu
ra
l I
ns
tit
ut
e
Texture gradient: a
monocular depth cue that
occurs because textures
become finer as they recede
in the distance
FIGURE 7.10 Texture Gradients
This field of tulips demonstrates how texture gradients show depth. As
with apparent size, as the tulips get farther away from us, their size on
the retinae becomes smaller. This helps create the experience of depth.
©
iS
to
ck
ph
ot
o.
co
m
/K
ar
en
M
as
si
er

195 Chapter 7: Depth and Size Perception
Atmospheric perspective: a
pictorial depth cue that arises
from the fact that objects in
the distance appear blurred
and tinged with blue
Shadows: a depth cue arising
because an object is in front
of its shadow; the angle of
the shadow can provide
information about how far
the object is in front of the
background
FIGURE 7.11 Atmospheric Perspective
Very distant objects tend to have a blue tinge to them. (a) The mountains in the Chilcotin region of British Columbia, Canada, take on a blue tinge. (b)
The mountains in the distance along the Blue Ridge Parkway in North Carolina also take on a distinct blue tinge. Because the atmosphere scatters
light, more distant objects will also have a blue tinge.
©
iS
to
ck
ph
ot
o.
co
m
/v
ix
te
rd
©
iS
to
ck
ph
ot
o.
co
m
/G
ib
so
nP
ic
tu
re
s
(a) (b)
the horizon have a clear blue tinge. The photograph in Figure 7.11b was taken along the
Blue Ridge Parkway in North Carolina showing how the parkway gets its name from
this depth cue.
If we were standing on the moon, which has no atmosphere, faraway objects would
appear neither blurred nor blue. But our atmosphere scatters light, and it scatters blue
light more than other wavelengths (which is why our sky appears blue). This is related
to how rainbows are formed. Blue light bends more easily than other wavelengths.
Indeed, that the atmosphere scatters light also plays a role in our depth perception.
The short or blue wavelengths of light are most easily scattered by the particles in the
atmosphere. In addition, the scattering occurs for all light, regardless of the direction
it comes from, giving us a blue sky. Thus, light coming from a distant object should
have some of its light scattered. That will have two effects on the light reaching our
eyes: (a) Because blue is scattered more, more distant objects should appear bluish, and
(b) because not all of the light is traveling in a straight line to us, more distant objects
should appear a bit fuzzy.
Shadows and Shading
Shadows may also enhance the perception of depth in images. Shadows provide a
depth cue because the object is in front of the shadow, and the angle of the shadow
can provide some information about how far the object is in front of the background.
Objects that are in shadow must be farther from the light than objects that are not in
shadow. In particular, on a curved surface, light falling on an object will create a pattern
of light and shadow. This gives us information about the relative depths of different
parts of a surface. The perception of shadow and the role it plays in determining depth
and shape is interesting because it depends on the assumption our perceptual system
makes about the direction of the sun. Shadows fall differently on an indentation, like
a hole, versus something that sticks up, like a hill (Figure 7.12). In Figure 7.13, we see
two different views of what really is the same image. In the first image, in Figure 7.13a,
the lighting appears to be coming from above and to the right. In the second image, in
Figure 7.13b, the picture is upside down, but we still assume the light is coming from
above and to the right. Note how this flipping of the picture changes our perception of

196 Sensation and Perception
FIGURE 7.12
Shadows
Shadows depend on assuming the
sun is in the sky. However, if the
object is an indentation, the shadow
will be toward the sun, whereas on
a bump or hill, the shadow is away
from the sun.
FIGURE 7.13 Shadows and Shading
We see two views of the same image. (a) The lighting appears to be coming from above and to the right.
(b) The picture is merely flipped, so that the light appears to be coming from below and to the left. Note how
this changes our perception of whether the circles in the rock art are bumps or indentations.
©
F
ro
ar
in
gu
s/
W
ik
im
ed
ia
C
om
m
on
s/

CC
-B
Y-
SA
-3
.0
/G
FD
L
©
F
ro
ar
in
gu
s/
W
ik
im
ed
ia
C
om
m
on
s/

CC
-B
Y-
SA
-3
.0
/G
FD
L
(a) (b)
whether the circles in the rock art are bumps or indentations. In Figure 7.13a, we see
the circles as indentations in the rock, but in Figure 7.13b, we see the circles as bumps.
In this way, shadows can cause different perceptions of the 3D structure of an object.
TEST YOUR KNOWLEDGE
1. Create a diagram that illustrates the different monocular depth cues.
2. Assess the relative contributions of relative size and familiar size as depth cues
and how they might work together to disambiguate perception.
Motion Cues
The monocular cues we have been discussing so far can be extracted from stationary
images. These are the pictorial cues to depth. However, once you start to move or
objects are set in motion, a number of other monocular cues to depth emerge; these are
the motion-based cues. From an observer’s point of view, objects moving at different
speeds can reveal information about relative distance. We discuss three motion cues to
depth here: (a) motion parallax, (b) accretion and deletion, and (c) optic flow. Each of
these motion cues is considered monocular because it can be seen with only one eye.
However, motion parallax and optic flow can create visual experiences that are similar
to those produced by the binocular cue of stereopsis.
Motion Parallax
Motion parallax is a monocular depth cue arising from the relative velocities of objects
moving across the retinae of a moving person. The term parallax refers to a change in
position. Thus, motion parallax is a depth cue caused by the movement of the viewer. In
this way, it is similar to the binocular cues, because binocular cues rely on two images
across space (one to each eye), whereas motion parallax depends on multiple images
across time (i.e., multiple images in a single eye). It is perhaps easier to think of what
motion parallax is by imagining yourself as a passenger in a car looking out the side win-
dow. The car is moving very fast down the highway. The objects very close to the window,
such as the small trees planted by the highway, seem to rush by. Beyond the small trees,
you can see a distant farmhouse. The farmhouse appears to move more slowly relative
to you in the car. You know that the trees and the farmhouse are standing still; you are
Motion parallax: a monocular
depth cue arising from the
relative velocities of objects
moving across the retinae of a
moving person
Deletion: the gradual
occlusion of a moving object
as it passes behind another
object
Accretion: the gradual
reappearance of a moving
object as it emerges from
behind another object

197 Chapter 7: Depth and Size Perception
the object that is moving. But your constant speed (of, say,
60 mph) creates the illusion that the trees are rushing by
but the farmhouse is not. Farther off, you see the tiny
image of a commercial jet airplane moving across the sky.
Although you know this plane is moving in excess of 500
mph, it does not look nearly as fast as the trees whizzing
close by. Thus, we can use this movement relative to the
viewer as a cue for depth. Those objects that appear to
move more quickly by are closer to us. Objects that appear
to move more slowly by are farther away. This is illustrated
in Figure 7.14. We also highly recommend that you view
the interactive animations of motion parallax on ISLE 7.3a.
A more technical description of motion parallax
involves considering one’s fixation point. If you are look-
ing directly at the farmhouse from your car window,
that farmhouse can be said to be your fixation point.
(Fixation points are also important when discussing ste-
reopsis.) We can then divide the world into points closer
to you than your fixation point, such as the trees, and
objects farther away from your fixation point, such as
the parked tractor beyond the farmhouse or the more
distant airplane. Objects closer to your position on the
highway will appear to move in a direction opposite to
your motion. Thus, even though the trees are station-
ary and it is the car you are in that is moving, the trees
appear to move very quickly in the opposite direction. In
contrast, objects farther than the point of fixation, such
as the parked tractor beyond the farmhouse, appear to
move in the same direction as you do. This movement
opposite to your own by near objects and in the same
direction as your own by far objects is motion parallax.
You can examine this effect in motion by watching the
interactive animation of motion parallax on ISLE 7.3a2.
Deletion and Accretion
Deletion is the gradual occlusion of a moving object as
it passes behind another object. Accretion is the grad-
ual reappearance of a moving object as it emerges from
behind another object. Think about being in a library.
You watch someone emerge from behind one bookshelf
and then disappear behind another bookshelf. When the
person first becomes visible, you note that the bookshelf
was not moving, but the person suddenly emerged. This
provides information about relative depth. The person
must be behind the shelf. Similarly, when the person is
“deleted” as she moves behind the next shelf, you again see from her movement the
relative positions of her and the furniture. We can formalize this in the following way:
The object that is being deleted and then later accreted is the object that is farther
away than the object we can see continuously, which is therefore closer. This can be
seen in Figure 7.16. You can also examine this effect in motion by watching the inter-
active animation of deletion and accretion on ISLE 7.3b.
MoveMove
Position 1
(a) (b)
Position 2 Position 1 Position 2
FIGURE 7.14 Motion Parallax
As you move, objects closer to you appear to move faster, whereas objects
farther away appear to move more slowly. This is because the closer
objects move a greater amount on your retinae from Position 1 to Position 2
than do the more distant objects. (a) Motion parallax to a close object.
(b) Motion parallax to a distant object. (c) You can experience motion
parallax a bit in an anaglyph stereogram. Put on your anaglyph glasses and
move your head from side to side while looking at the image. You will see
the finger of the front figure move more than the faces in the background.
ISLE 7.3
Motion Depth Cues
(c)

198 Sensation and Perception
FIGURE 7.16 Deletion and Accretion
As we watch an object move relative to another object, we judge the
object that disappears (deletion) and reappears (accretion) as being
farther away from us than the object that is continually visible.
Time 1 Time 2 Time 3 Time 4
Optic flow: a motion depth
cue that involves the relative
motion of objects as an
observer moves forward or
backward in a scene
Optic Flow
Optic flow is the motion depth cue that refers to the relative motions of objects as
the observer moves forward or backward in a scene. Optic flow is related to motion
parallax. However, optic flow refers to our perception of objects as we move forward
or backward in a scene (Gibson, 1950) (Figure 7.17). Imagine now that you are driv-
ing down a straight country road. As you move forward, the world rushes by you
in the opposite direction. In front of you, however, the world is still coming toward
you and getting larger as it does. We can determine depth from optic flow because
faraway objects appear to move more slowly relative to closer objects, which appear
to rush toward us more quickly. Indeed, extremely large faraway objects may appear
essentially fixed in position. And the objects that are most close in front of you rush
by you at high speeds. Optic flow is often used to convey depth information in mov-
ies. Think of the stars expanding outward from a point when the Millennium Falcon
enters hyperspace in a Star Wars movie. The moving stars create an optic flow and
it can help you feel as if you are moving with the ship. Car chases in movies also use
optic flow to get you to feel as if you are in the car chasing the bad guys at high speed.
You can also examine an interactive animation of optic flow on ISLE 7.3c.
Now imagine you are driving a car on a winding country road. The motion cues
for depth are a combination of optic flow, relative movement coming directly at you,
and motion parallax, relative movement at a 90-degree angle from you. Add in hills for
Focus of
expansion
FIGURE 7.17 Optic Flow
Consider being in the driver’s seat and
seeing the view in front of you. As you
move forward, the world moves toward
you and then disappears behind you.
Your fixation point remains constant in
the distance, but objects flow toward
you and spread out as they do, relative
to your position. The arrows in the figure
on the right side represent this relative
movement as you drive down the road.
iS
to
ck
ph
ot
o.
co
m
/k
av
ra
m
FIGURE 7.15 Aerial Views of New York
An American Airlines plane flies over a highway September 13, 2009,
in the Queens borough of New York City.
M
ar
io
T
am
a/
G
et
ty
Im
ag
es
N
ew
s/
G
et
ty
Im
ag
es

199 Chapter 7: Depth and Size Perception
up-and-down motion, and even more depth cues become available. The amazing thing
about our visual system is that it computes all these relative distances and speeds fast
enough for us to make sense of the world, even at 65 mph.
TEST YOUR KNOWLEDGE
1. Compose a passage for a friend describing how motion parallax works to
contribute to the perception of depth.
2. Relate deletion and accretion as depth cues to occlusion.
OCULOMOTOR CUES
7.2 Compare the different oculomotor curves in their effectiveness as depth cues.
Rest your index finger on the tip of your nose and look at your finger. Then, without
moving your head or the direction of your gaze, adjust your focus so that you are look-
ing at an object farther away, such as the wall of the room in which you are reading.
As you adjust your eyes, you can feel the movements involved in your focus. These
adjustments that your eyes make as they look from objects near to objects far away or
from objects far away to objects closer can serve as cues to depth in and of themselves.
The movements involved come from two sets of muscles, one that controls the shape of
the lens (accommodation) and one that controls the rotation of the eyes relative to each
other (vergence). We look at each in turn.
Accommodation
Accommodation is the process of adjusting the lens of the eye so that you can see
both near and far objects clearly. We discussed this process in Chapter 3. To focus on
a more distant object, we relax the ciliary muscles that contract the lens, and to focus
on an object close to our eyes, we contract the ciliary muscles. These contractions and
relaxations take place automatically as we change our focus. However, we can sense
these muscle movements. And it is the sensing of these movements that could give
us information about depth. However, the narrowness of the pupil during the day
makes accommodation imprecise in day vision, and the poor acuity at night makes
accommodation imprecise then as well. So, accommodation probably does not play
a large role in depth perception except as related to vergence (Kapoula, Bernotas, &
Haslwanter, 1999).
Vergence (or Convergence)
Vergence (also known as convergence) occurs when the eyes rotate inward to see a near
object and then bend outward (diverge) when we look at a more distant object. When
you are looking at your finger resting against your nose, your eyes are rotated inward
to place the image of this very close object on the foveae of both eyes. Essentially, to see
your finger, your gaze must become “crossed.” When you shift your gaze to the window
or wall beyond your eyes, your eyes rotate outward from each other (Figure 7.18).
As with accommodation, this process is automatic, but our brain can sense the move-
ments, and this feedback gives us information about the relative distance of objects.
Convergence is probably a more useful cue than accommodation, and it can provide the
Vergence: the inward bending
of the eyes when looking at
closer objects

200 Sensation and Perception
visual system with reliable depth information to about 2
meters (m) in distance, at which point there is no appre-
ciable difference in eye angle (Schachar, 2006). You can
explore an interactive illustration of accommodation
and vergence on ISLE 7.4 and 7.5, respectively.
TEST YOUR KNOWLEDGE
1. What is accommodation, and how might it be used
to infer depth?
2. Describe vergence as a depth cue and assess
its relative contribution to depth perception with
accommodation.
BINOCULAR
CUES TO DEPTH
7.3
Summarize the principle of stereopsis and
how it applies to human depth perception.
At the beginning of the chapter, we introduced jumping
spiders, amazing creatures with eight eyes (and eight
legs). Human beings are not fortunate enough to have
eyes at both the front and the back of our heads. But
we do have two eyes, located adjacent to each other at
the front of our heads. Because our eyes are next to each
other at the front of our heads, each eye sees mostly the
same objects in the visual world as does the other eye
but from a slightly different angle (review Figure 3.5).
This contrasts with animals such as horses, which have
eyes on the opposite sides of their heads, such that there
is almost no overlap in what each eye sees. This gives
horses a greater ability to see completely around them.
Horses essentially have a panoramic view. However,
because the fields of view of the eyes do not overlap,
horses cannot take advantage of the binocular depth cues that we do. That we humans
have two eyes that see the world from slightly different angles provides us with an
important cue to depth. The reason our two eyes both look in the same ways is that we
pick up information about depth perception from the overlap of the two visual fields.
In the area where both eyes see the same part of the world, we have binocular vision.
We define stereopsis as the sense of depth that we perceive from the visual system’s
processing of the comparison of the two different images from each retina.
Focus your gaze on some nearby object. If there is nothing readily available, hold
one hand out at arm’s length and look at your own index finger. First look at that index
finger with your left eye and then with your right eye. You get an illusion that your hand
has shifted somewhat as you go back and forth from one eye to the other. This quick
test allows us to see the slightly different perspective on the world that each eye gives
us. If you have an object to look at just beyond your index finger, you can see that the
object may appear closer to your index finger through one eye than through the other
(a) (b) (c)
33˚ 3.5˚ <1˚ 10 cm from eyes 1 m from eyes 4 m or more from eyes FIGURE 7.18 Vergence Your eyes cross to focus on a nearby object. People can “feel” this movement, and that gives them information that objects are close by. When looking at an object farther away (more than 4 m), there is no vergence at all. This, too, provides information about depth. ISLE 7.4 Accommodation ISLE 7.5 Vergence Stereopsis: the sense of depth we perceive from the visual system’s processing of the comparison of the two different images from each retina 201 Chapter 7: Depth and Size Perception (Figure 7.19). What is important to binocular vision is this area of overlap. We see the same objects in the area of overlap, but it is not the same image. Binocular vision involves comparing the two images. The sense in which stereopsis changes the experience of depth perception is apparent to anyone who has seen a 3D movie, which recently have become quite popular again (this technology also relies on vergence). In 2017, movies as varied as Star Wars Episode 8: The Last Jedi, The Lego Batman Movie, and Spider-Man: Homecoming were available in theaters in 3D format. The sense of objects coming out of the screen toward you in the audience can often distract from plot development and good acting, but it is a perceptually unmistak- able phenomenon. The sense of Spider-Man flying out of the screen toward you in three dimensions is a perceptually very different expe- rience than watching the same movie in standard format (Figure 7.20). We see both 2D and 3D movies on flat screens, whether they are movie screens or smaller television screens. But the 3D glasses one wears while watching a 3D film create a situation in which a slightly different image is sent to each eye. This allows the use of bin- ocular cues to extract depth from the images. To see how 3D movies work, refer to Figure 7.21. You can also examine a demonstration of 3D movies on ISLE 7.6. Binocular Disparity Binocular disparity arises because our two eyes are in different locations in our head and therefore have slightly different views of the world. The explanation of how disparity allows us to extract depth information is quite complex. So, read through this section slowly, make sure you understand the diagrams, and review it until you know it well. After you are certain that you know it well, draw yourself a diagram of how the two eyes perceive through stereopsis and explain it to someone who is not taking your sensation and perception class. Only after you have done that can you be sure that you understand the explanation of stereopsis. You can see a demon- stration of binocular disparity on ISLE 7.7. Line of gaze from left eye Line of gaze from right eye View from left eyeView from right eye FIGURE 7.19 Binocular Disparity Each eye has a slightly different view of the world. If you hold your finger out at arm’s length and then look at it alternately with your left eye only and then your right eye only, the image of your finger relative to the world behind it will shift somewhat. This is binocular disparity, which helps provide the basis for the determination of depth. ISLE 7.6 Stereopsis ISLE 7.7 Binocular Disparity Binocular disparity: a binocular depth cue because our two eyes are in different locations in our head and therefore have slightly different views of the worldFIGURE 7.20 Spider-Man: Homecoming Pi ct or ia l P re ss L td / A la m y St oc k Ph ot o 202 Sensation and Perception FIGURE 7.21 Three-Dimensional Movies (a) A traditional 2D movie. In these movies, each eye sees the same information, and depth is inferred from monocular and motion cues. (b) A 3D movie. In this situation, each eye receives a different image, allowing stereoscopic vision to occur. Left eye image Right eye image Left eye image Right eye image Perception = 2D (a) Same image to left and right eyes (b) Different images to left and right eyes Perception = 3D 2D image on flat screen 3D information on flat screen 3D glasses Corresponding points: refers to a situation in which a point on the left retina and a point on the right retina would coincide if the two retinae were superimposed Noncorresponding points: refers to a situation in which a point on the left retina and a point on the right retina would not coincide if the two retinae were superimposed Horopter: the region in space where the two images from an object fall on corresponding locations on the two retinae Panum’s area of fusion: the region of small disparity around the horopter where the two images can be fused into a single perception Diplopia: double images, or seeing two copies of the same image; usually results from the images of an object having too much disparity to lead to fusion Corresponding and Noncorresponding Points Look around you and find some reference points to use when following this argu- ment. If you are in a room in your house, a dorm room, or a library, there is likely a lamp nearby. Position yourself so that you are about 10 feet (3 m) from that lamp. Hold your thumb out in front of you. Notice how you can use convergence and accommodation to change your focus from looking at your thumb to looking at the lamp. When you focus on the lamp, you may also notice that you see a double image of your thumb. When you focus on your thumb, you may notice two images of the lamp in the background. So, let us see what is going on with your visual system as you do this. For this next discussion, imagine taking out your two retinae and overlapping them such that the two foveae are on top of each other. Corresponding points refers to a sit- uation in which a point on the left retina and a point on the right retina would coincide if the two retinae were superimposed; that is, if you put a pin straight through these two overlapping retinae, it goes through corresponding points. In contrast, noncorrespond- ing points refers to a situation in which a point on the left retina and a point on the right retina would not coincide if the two retinae were superimposed. For example, when looking at your thumb with your hand held out at arm’s length, the image of your thumb falls on corresponding points on your left and right retinae. The lamp beyond it does not. This is why you have the illusion of seeing two lamps (see Figure 7.22). Now imagine a semicircle in front of you, with all the points at the same distance from your eyes as your thumb. This imaginary line is the horopter. Thus, if you stretch out your other arm along- side the one you are fixated on, your other thumb is not your fixation point, but lies along the horopter. Technically, the horopter is the region in space where the two images from an object fall on corresponding locations on the two retinae. If you switch your focus 203 Chapter 7: Depth and Size Perception from your index finger to the lamp, you now have established a new horopter. Because you are now fixating on the lamp, the horopter is now an imaginary semicircle of points that is the same distance from your eyes as is the lamp. Please see Figure 7.22 and go to ISLE 7.8 for a demonstration of this phenomenon. Objects that lie along the horopter are perceived as single unified objects when viewed with both eyes. We can also fuse into a single object the images from the left and right eyes for objects that fall inside Panum’s area of fusion (see ISLE 7.8b). Inside Panum’s area, including the horopter, we see images as singles. Outside Panum’s area, either in closer or farther, we see double images. As we move from the horopter toward our eyes, we are looking at objects closer and closer to ourselves. At some point, we lose the perception of these objects as being single unified objects and we see instead two images of the same object. Thus, when looking at the lamp 3 m away, the image of your index finger appears as a double. This double vision is known as diplopia. Examine Figure 7.23a. We can see schematically the retinae of the left and right eyes. In the diagram, the eyes are fixated on Object A, depicted in red in the diagram. The image of A falls on Points A and Aʹ on the left and right retinae, respec- tively. Because we are fixated on Image A, it defines the horop- ter for our visual world. Object B, for example, lies along the horopter. Thus, Object B falls along Points B and Bʹ on the retinae. We perceive Object B as being in the same plane or the same distance from us as Object A. Now consider Object C. Object C lies closer to us than does Object A and falls on Points C and Eʹ on the retinae. Notice how Point C is closer to B than Eʹ is to Bʹ. Because it does not lie along the horopter, its image falls on noncorresponding points. Object D lies farther from us than Object A and falls on Points D and Fʹ on the ret- inae. Again, notice how Point D is closer to Point A than Fʹ is to Aʹ. Because it does not lie along the horopter, its image falls on noncorresponding points. Points that are nearer to us than the horopter have crossed disparity, and points farther away have uncrossed disparity. You can see this in the diagram. Bear with this example, as we introduce a little more ter- minology necessary to understanding binocular vision. Yes, the visual world does get much more complicated because of binocular vision. Consider Object C in Figure 7.23b. Object C is closer to the observer than the horopter line. Its image falls on Points C and Eʹ on the retinae. Look closely at these points. You see Object C to the right of the fixated Object A with the left eye, but to the left of the fixated Object A with the right eye. Because of this differential position along the retinae, we call this crossed disparity. Crossed disparity refers to the direction of disparity for objects closer to the viewer than the horopter (the image in the left eye is to the right of the image of the object in the right eye). This disparity is often given a positive sign. You can also see this interactively on ISLE 7.8c. Fovea Horopter Horopter FIGURE 7.22 The Horopter and Corresponding Points When we look at an object, that object defines our horopter—all points equally distant from us form the horopter. Technically, the horopter is the region in space where the two images from an object fall on corresponding locations on the two retinae. Points that are closer to us than the horopter have crossed disparity, and points farther away have uncrossed disparity. ISLE 7.8 The Construction of Visual Depth With Binocular Disparity Crossed disparity: the direction of disparity for objects closer to the viewer than the horopter (the image in the left eye is to the right of the image of the object in the right eye) 204 Sensation and Perception Next consider Object D in Figure 7.23b. Object D is farther from the viewer than the horopter line. The image of Object D falls on Points D and Fʹ of their respective retinae. Note that the image of Object D falls to the left of Object A (the fixated object) on the left retina, but to the right of Object A on the right retina. Because of this posi- tioning along the retinae, we call this uncrossed disparity. Uncrossed disparity is the name given to the direction of disparity for objects that are farther from the viewer than the horopter (the image of the object in the left eye is to the left of the position of the image of the object in the right eye). This type of disparity is often given as a negative value. Finally, consider Object A itself. Its image falls on corresponding points on the left and right retinae. Similarly, Object B, though not the focus of our fixation, also falls along the horopter, so its image falls on corresponding points on the left and right reti- nae. Points along the horopter are said to have zero disparity. Zero disparity means that retinal images fall along corresponding points. So, now let us examine Figure 7.24. We now consider the magnitude of reti- nal disparity between the two images. We can see here that the amount or magni- tude of disparity increases as the distance of an object from the horopter increases. Consider Figure 7.24a. This figure shows two objects with crossed disparity, that FIGURE 7.23 Corresponding and Noncorresponding Points and the Horopter (a) Corresponding points are for objects that fall on the horopter. (b) Crossed disparity is for objects (C) that are in front of the horopter and uncrossed disparity is for objects (D) that are behind the horopter. Horopter Fixation point AB D C Left eye D A B C F´ A´ B´ E´ Right eye (a) Horopter Fixation point A D C Left eye C A D E´A´F´ Right eye (b) Uncrossed disparity: the direction of disparity for objects that are farther from the viewer than the horopter (the image of the object in the left eye is to the left of the position of the image of the object in the right eye) Zero disparity: the situation in which retinal images fall along corresponding points, which means that the object is along the horopter 205 Chapter 7: Depth and Size Perception is, two objects that are closer to the viewer than the object being fixated on. You can see that Object C is closer to the viewer than Object B. Therefore, Object C has a greater crossed disparity than Object B. Now consider Figure 7.24b. This figure shows two objects with an uncrossed disparity, that is, two objects that are farther from the viewer than the object being fixated on. You can see that Object D is farther from the viewer than Object E. Therefore, Object D has a greater uncrossed disparity than Object E. We now have the cue people use to determine relative distance of objects by compar- ing the relative positions of the images of objects on the two eyes. The degree of retinal disparity gives us information about objects near and far. Indeed, retinal disparity can give us information about depth distances as small as 4 mm at a distance of 5 m. It is also useful at distances of up to 200 m (roughly an eighth of a mile) (Howard & Rogers, 2002). Horopter Fixation point Left eye view Fovea C B A F E D Fovea Right eye view Left eye view Right eye view Larger crossed disparity Smaller crossed disparity (a) (b) Horopter Fixation point Fovea Fovea Smaller uncrossed disparity Larger uncrossed disparity FIGURE 7.24 Magnitude of Binocular Disparity The farther away an object is from the horopter, the larger the disparity will be. (a) This figure shows two objects with crossed disparity, that is, two objects that are closer to the viewer than the object being fixated on. You can see that Object C is closer to the viewer than Object B. Therefore, Object C has a greater crossed disparity than Object B. (b) This figure shows two objects with an uncrossed disparity, that is, two objects that are farther from the viewer than the object being fixated on. You can see that Object D is farther from the viewer than Object E. Therefore, Object D has a greater uncrossed disparity than Object E. 206 Sensation and Perception Think again about 3D movies. The plane of the screen becomes the horopter. The manipulation of images to the left and right eyes allows the filmmaker to create images that go to noncorresponding points along the retinae of the left and right eyes. If the filmmaker creates an image with uncrossed disparity, we see that image as being behind the plane of the screen. If the filmmaker creates an image with crossed disparity, we see that object as being in front of the screen. Indeed, if we perceive an object with increas- ing crossed disparity, this object appears to come out of the screen toward us. Think of the shark in the 1983 movie Jaws 3-D. The approaching image of the deadly shark is caused by increasing crossed disparity. If you have not seen this movie, don’t bother. It is awful, but the image of a giant shark coming out of the screen to attack you is a powerful illustration of stereopsis. In everyday vision, our brain automatically uses binocular disparity as a cue for depth. Although the monocular cues provide good estimates of depth and distance, binocu- lar disparity provides strong subjective cues. When watching a regular movie, we have no problem judging distances between characters and objects. However, the binocular cues available in 3D movies create a strong experiential boost to our perception of three dimensions. Seeing objects move into and out of the screen, much as they do in real life, reinforces the importance of binocular disparity in creating the perception of a 3D world. The Correspondence Problem As we just described, the brain uses crossed and uncrossed disparity to determine if objects are in front of or behind the horopter and the magnitude of disparity to deter- mine how far the object is from the horopter. But there is an important task our visual system must accomplish that must be examined. How does our visual system know which image in the left eye matches up with which image in the right eye? In simple scenes, this may be relatively obvious. Because we know a lot about objects, in some cases, matching the image in the left and right eyes should be easy. For example, when looking at the lamp, that the “finger-shaped object” about a meter away from your eyes is your index finger in both your left and right eyes is not surprising. But in many complex scenes, matching images to the left and right eyes may be more difficult. Real scenes often involve complex textures, similar objects in motion, and many other vari- ations that may cloud an easy linking of an image in one eye to an image in the other. Imagine watching a raft of ducks swimming by. Some ducks are closer than others, but they all look identical to the untrained eye. How do our eyes match up the cor- rect image of ducks in the left and right eyes? What happens if we match the wrong ducks together across eyes? This is the correspondence problem. The correspondence problem (depth perception) is the problem of determining which image in one eye matches the correct image in the other eye. How our visual system solves the correspon- dence problem has been an area of some fascinating research, which we discuss shortly. But to understand how our visual system solves the correspondence problem, we must describe the techniques researchers have used to investigate the issue. This means we will have to take a quick look at the nature of 3D images called stereograms so that we can discuss how random-dot stereograms help us understand how our brain solves the correspondence problem. TEST YOUR KNOWLEDGE 1. What is the difference between crossed and uncrossed disparity? Which corresponds to points closer than the horopter, and which corresponds to points farther away than the horopter? 2. How does binocular disparity arise? What aspects of the placement of the eyes are important? Correspondence problem (depth perception): the problem of determining which image in one eye matches the correct image in the other eye 207 Chapter 7: Depth and Size Perception STEREOGRAMS 7.4 Sketch the correspondence problem and how it must be solved for stereopsis. Charles Wheatstone (1802–1875) was an English inventor during the 19th century. Among his inventions were a concertina (a small handheld accordion-like instrument), devices necessary for the creation of the first telegraph networks, and his famous Wheatstone stereoscope, invented before the first photographs. In the Wheatstone ste- reoscope, you looked at the two mirrors and one image was presented in each mirror giving the two images needed to create disparity (Figure 7.25a). This was a cumbersome instrument, and it was quickly replaced by the Brewster stereoscope, which became very popular in the 19th century. The Brewster stereoscope is a small instrument (see Figure 7.25b) that presents one image to one eye and a second image to the other eye using prisms to help the viewer. The pictures presented to each eye are images slightly offset from each other in order to replicate the phenomena of crossed and uncrossed disparity. When one looks through a stereoscope, one sees a single 3D image of the scene. Originally, Wheatstone used drawings for his stereograms, as photography was still a year away from being invented and thereafter was expensive, but with the advent of photography, stereoscopic photographs happened very quickly. It is fairly easy to take stereographic pictures. A photographic stereogram can be made by taking a photograph of a scene. First, select a “still-life” type scene, perhaps a bowl of fruit. Take a photograph and then move the camera 4 inches (6 cm) to the right and take a second photograph. Keep in mind that the first photograph is the left-eye photo and the second photograph is the right-eye photograph. Print out the photographs and place them in an object like the stereoscope above. Such an image is presented in Figure 7.26. When we look at the stereogram through a stereoscope, one image goes only to the left eye, and one image goes only to the right eye. Because the images are so similar, they are combined by the visual system, and the result is the illusion of depth. Some people can examine stereograms without the use of a stereoscope, a skill known as free fusion. The way to do so is to control the convergence movements of one’s eyes. If one relaxes one’s gaze and imagines that one is looking at a distant point, the left and right eyes may diverge sufficiently such that the left eye is seeing only the left-eye image, and the right eye is seeing only the right-eye image. If this is difficult, you may be able to free-fuse the images by crossing your eyes. This means maximally converging your eyes as if you were looking at an object very close. This usually involves reversing the pho- tos, because when you cross your eyes, you will be seeing the left-eye image with your FIGURE 7.26 An Old- Fashioned Stereogram FIGURE 7.25 Stereoscopes (a) A Wheatstone stereoscope. The viewer looks at the mirrors and the images for each eye are off to the side. (b) The Brewster stereoscope. This simple device projects one image to each eye. If there is disparity in the images, a 3D image will jump out at the viewer. (c) A Brewster stereoscope as an anaglyph stereogram. Use your anaglyph glasses. © Brand X Pictures/ Stockbyte/Thinkstock (a) (b) (c) By Creator:Life G roups [CC0], via W ikim edia Com m ons 208 Sensation and Perception right eye and the right-eye image with your left eye. So, if you copy, cut out, and reverse the photos as usually depicted, you can see the correct depth relationship with a crossed-eye approach. In both forms of free fusion, one is usually left with the odd perception of see- ing three pictures. You can still see each of the original pictures, but you now see a third picture in the middle from the combination of your two eyes. It is this third (and essentially illusory) picture in which you can see the depth relations (see ISLE 7.9 for some really fun examples of stereograms). The anaglyph is another form of stereogram and the technique generally used in older 3D movies. Like traditional stereograms, anaglyphs are made by taking two pho- tographs of a scene from cameras separated by about 6 cm. However, one photograph is then printed in a shade of one color, such as blue, whereas the other photograph is printed in a shade of another color, such as red. The two photographs are then integrated into a common image. The common image looks a bit fuzzy under normal viewing. However, when viewed through special color-coded anaglyph glasses, each of the two images goes to one eye, allowing for the stereoscopic image to emerge. The current convention is for the red lens to cover the left eye, and the cyan (blue) lens to cover the right eye. If you look at Figure 7.27 first without anaglyph glasses, you see a fuzzy image. But then put on a pair of anaglyph glasses and look at Figure 7.27 again. As with the stereograms, your left eye now sees one image and your right eye another image, and your visual cortex extracts the 3D information from the photograph. In one photograph, we see two students studying in their science classroom, but the anaglyph glasses give us a stunning illusion of depth. Similarly, the masks seem to pop out of the book when you look at them through the glasses. Make sure to keep your anaglyph glasses handy. There are other anaglyphs scattered throughout the text. Go to ISLE 7.9 if you want to see more of these. Random-Dot Stereograms Earlier, we introduced the correspondence problem, that is, how the visual system determines which object’s image in one eye matches the same object’s image in the other eye. Here we explain how stereograms can be used to study the correspondence problem and potentially derive some solutions. How the visual system is able to solve the correspondence problem is quite complex computationally. The specifics of this computation were the subject of a debate that took place about the nature of stereop- sis. On one side was the view that the visual system must bring to bear knowledge of objects and use that knowledge to match images. The other view was that the visual system matches the left and right images before bringing knowledge of the objects to bear on the situation. To distinguish these hypotheses, Hungarian American vision researcher Béla Julesz invented random-dot stereograms (Julesz, 1971). Random-dot stereograms are stereo- grams in which the images consist of a randomly arranged set of black and white dots. The left-eye and right-eye images are arranged identically, except that a portion of the dots is moved to the left or the right in one of the images to create either a crossed or an uncrossed disparity. This creates the experience that part of the image is either in front ISLE 7.9 Anaglyph Stereograms FIGURE 7.27 Anaglyph This is another form of stereogram and the technique generally used in recent 3D movies. Use a pair of anaglyph glasses to look at these images. (a) (b) Random-dot stereograms: stereograms in which the images consist of a randomly arranged set of black and white dots, with the left-eye and right-eye images arranged identically except that some of the dots are moved to the left or the right in one of the images, creating either a crossed or an uncrossed disparity ISLE 7.10 Random-Dot Stereograms 209 Chapter 7: Depth and Size Perception of or behind the rest of the dots. Such a stereogram is seen in Figure 7.28. Figure 7.29 shows random-dot ste- reograms as anaglyphs. A graphic on how random-dot stereograms are made can be seen in Figure 7.30. You can explore an interactive illustration of random-dot stereograms on ISLE 7.10. When you view a random-dot stereogram without a stereoscope or without free fusing, all you see is a grid of white and black dots. There is no shape appar- ent other than this uninteresting field of dots. Thus, if we can extract depth information from such images, it must come from depth perception processes that pre- cede object recognition, because there are simply no objects to perceive in such figures until after stereopsis has occurred. Thus, the argument Julesz (1971) made is the following. Correspondence between points in the left image and right image is necessary for the perception of binocular disparity. If object recognition is necessary for matching correspondence, then random-dot stereograms will not result in a 3D perception. However, if correspon- dence matching occurs before object recognition, then it should be possible for people to extract binocular depth cues from random-dot stereograms. Inspect Figure 7.28 through a stereoscope or Figure 7.29 or ISLE 7.10 with a pair of 3D glasses. What do you see? Do any patterns jump out of the page at you? The majority of people with nor- mal stereopsis will see a square made up of many random dots jump out in front of the page. If you are free fusing by crossing your eyes, the figure will appear behind the plane of the page. Thus, because people do see depth information in random-dot ste- reograms, we know that correspondence precedes object recognition. The Anatomy and Physiology of Binocular Perception The next question we can ask about stereopsis is how it is achieved in the human visual cortex. We already touched on the answer to this question in Chapter 4. We review these issues and expand on them here. As discussed in Chapter 4, there are cells sensi- tive to binocular disparity in V4 and also in V1 of the occipital cortex (Hubel & Wiesel, 1962). These binocular cells have two receptive fields, one for each eye. These cells are also usually similar with respect to their preferred orientation and motion sensitivity, suggesting that the main function of these cells is to match the images coming to each eye. Moreover, many binocular cells in the cortex respond optimally when the images are on corresponding points on each retina. Interestingly, there are also binocular cells that respond best to varying degrees of disparity, that is, when similar images lie in front FIGURE 7.28 Random-Dot Stereograms When these images are viewed through a stereograph, you can see patterns floating in front of the black and white dots. FIGURE 7.29 Random-Dot Stereograms as Anaglyphs Use a pair of anaglyph glasses to examine the image. You should see patterns floating in front of the pattern of black and white dots. 1 2 1 2 1 2 2 1 2 1 1 2 1 1 2 1 2 2 2 2 2 1 1 2 1 2 1 1 1 2 1 2 2 2 2 1 2 1 2 1 1 2 1 1 2 1 2 2 2 1 2 1 2 2 2 2 1 2 1 2 1 2 1 1 1 2 1 2 1 1 1 2 2 1 2 2 2 1 2 2 1 2 1 2 1 2 1 1 1 1 2 1 2 1 1 1 1 2 1 2 1 2 1 2 1 2 2 1 2 1 1 2 1 1 2 1 2 2 2 2 2 1 1 2 1 2 1 1 1 2 1 2 2 2 1 2 2 1 2 1 1 2 1 2 1 2 1 2 2 1 2 1 2 2 2 1 1 2 2 1 2 1 2 2 2 1 2 2 1 2 1 2 1 1 2 1 1 2 1 1 1 2 2 1 2 2 2 1 2 2 1 2 1 2 1 2 1 1 1 1 FIGURE 7.30 How Random-Dot Stereograms Are Made To make a random-dot stereogram, you make a random grid of black and white dots. Each number in the figure represents either a black (1) or white (2) dot. You then copy the image to make two such images. But in the second image, you shift a central section of the first image to the right or left. Thus, the same pattern is represented in each image, but part of it is shifted. When we look at this through a stereograph, the shifted part will appear either in front of or behind of the rest of the dots, depending on which direction it was shifted. Binocular cells: cells with two receptive fields, one for each eye; their main function is to match the images coming to each eye 210 Sensation and Perception of or behind the horopter (Barlow, Blakemore, & Pettigrew, 1967). This arrangement is depicted in Figure 7.31. Indeed, different binocular cells are tuned to different dispari- ties. For example, a neuron may be tuned to crossed disparity of a particular magnitude, whereas another cell might be tuned to an uncrossed disparity of another magnitude. Moreover, such disparity-tuned cells are found throughout the visual cortex, including along both the ventral pathway (what) and the dorsal pathway (where and when). For both systems, depth information can play an important role in visual perception (Parker, 2007). Developmental Issues in Stereopsis Are human beings born with stereopsis, or is it a visual skill that requires experience in the environment to learn? If it does develop, what are the perceptual and physiological processes that allow infants to learn stereopsis? Research suggests that newborn infants are blind to binocular depth information and continue to show no stereopsis until the age of about 4 months. At about 4 months of age, stereopsis develops rapidly in human infants (Teller, 1983). Indeed, a number of studies have found that 3-month-old infants do not detect disparity at all, but by 5 months of age, infants perceive depth from dis- parity as well as normal adults (Birch, 1993). Testing infants at this age is tricky because they cannot make verbal responses, nor are they yet in control of all their muscular movements. Thus, researchers must be clever in designing tasks on which infants can provide measurable responses. It turns out that most infants like novel stimuli. When presented with an image they have seen before or a completely new one, infants will direct their gaze toward the new stimulus. This is often called a novelty preference. Gaze direction and novelty preference can be measured by researchers. Thus, we can habit- uate an infant to a 2D stimulus. Once the stimulus is no longer novel, we can show the infant an alternative stimulus that has the same structure in two dimensions but has a 3D interpretation if the infant can use binocular cues. Thus, if the infant prefers the 3D image to the 2D image, it demon- strates stereopsis, as only with the 3D information is the image novel. Preferential looking tasks show that infants start becoming sensi- tive to disparity at about 5 months of age (Birch, 1993). One explanation for this phe- nomenon is that binocular cells in the cortex are not mature and not yet functioning. (It also could be infants’ poor acuity and lack of vergence.) The current research suggests that the area in the brain that is not yet mature is found not in V1 but in higher areas of the occipital cortex. Chino, Smith, Fixation point Binocular neurons in striate cortex (a) (b) (c) FIGURE 7.31 Binocular Cell Receptive Fields These are simplified illustrations of what a receptive field for disparity-tuned neurons in the occipital cortex look like. (a) Both the red cell and the blue cell are not responding, as the only object is the fixation point, to which disparity-tuned neurons do not respond. (b) There is an object closer than the fixation point. A crossed disparity neuron (red) is firing in response to this object. (c) There is an object farther away than the fixation point. A blue uncrossed disparity neuron is firing in response to this object. 211 Chapter 7: Depth and Size Perception Hatta, and Cheng (1997), for example, found that binocular cells of 1-week-old monkeys were responding to disparity in the same way that binocular cells of older monkeys were doing. Ocular dominance columns also appear to be mature at birth. At this point, it is not clear where the locus of origin in the brain for stereopsis to emerge is, but some evidence points to area V2 (Zheng et al., 2007). Nonetheless, as of yet, there is no ready explanation for the sudden development of stereopsis at around 4 months of age. TEST YOUR KNOWLEDGE 1. What is a random-dot stereogram? How is it made? What issue does its study address? 2. What do random-dot stereograms tell us about how the visual system solves the correspondence problem? SIZE PERCEPTION 7.5 Explain the concept of size perception and the inferential nature of its determination. Size and depth are deeply intertwined. Think of watching a professional basketball game from the seats at the top of the coliseum. One knows the players are taller than normal people, but at that distance, it is hard to perceptually distinguish just how tall they are. Contrast this with the experience of finding yourself next to Shaquille O’Neal at the deli counter of your supermarket. At close range, his extreme tallness is appar- ent, whereas at a distance, we use familiar size cues to perceive people, any people, and thus, the basketball player’s height is underestimated. This is similar to the experience we have looking at the ground from airplanes. Intellectually, we understand that we are 2 miles above the surface of the earth and the cars and houses we see are really quite large. But at that vertical distance, our sense of size is problematic, and we perceptually experience what look like toy cars and houses. In both of these cases, we misjudge size because we are at unfamiliar distances. We are more familiar with interacting with peo- ple, houses, and cars at closer distances. At these familiar distances, we can scale size better. Indeed, at most distances, we can judge the sizes of objects quite well, despite the changes in size these objects make on our retinae as they approach us or recede from us. Think of watching your friends leave your house. As you watch their car speed away down the street, you do not see your friends’ car getting smaller. You perceive the car at the same size but moving farther away from you. This leads us to the concept of size–distance invariance. Size–distance invariance refers to the relation between perceived size and per- ceived distance and simply states that the perceived size of an object depends on its perceived distance, and the perceived distance of an object may depend on its perceived size. In a classic experiment, Holway and Boring (1941) looked at this relation. (College students may not make jokes about the second author’s name. He was a famous psychologist.) They placed disks at various distances from the observers in such a way that the disks all took up the same amount of space on the observers’ retinae. If there were no depth cues, the observers judged the objects to be of the same size, a natural judgment given that the objects were equivalent on the retinae. But when depth cues were provided, the observers correctly distinguished the smaller objects up close from the larger objects farther away. That is, as soon Size–distance invariance: the relation between perceived size and perceived distance, whereby the perceived size of an object depends on its perceived distance, and the perceived distance of an object may depend on its perceived size 212 Sensation and Perception as distance was clear, the observers’ visual systems used the size–distance invariance principle to scale the sizes of the objects consistent with their distance. Important to studying size–distance invariance is the concept of a visual angle. A visual angle is the angle of an object relative to an individual’s eyes. That is, if we drew lines from the top and bottom of an object and extended those lines to your eyes, we would have the visual angle of that object (Figure 7.32). Smaller objects close up can have the same visual angles as larger objects farther away. Classic examples of this are the sun and the moon. Each appears approximately the same size in the sky and makes a similar visual angle to our eyes (by sheer coinci- dence). However, the moon is much smaller than the sun, and the moon is also much closer to Earth than the sun is. Similarly, an index finger held at arm’s length also makes approximately the same visual angle as do the sun and the moon (although, again, your index finger is obviously much smaller and closer than either the sun or the moon). We perceive the index finger as smaller than these faraway objects because of size–distance invariance. However, we cannot perceptually deter- mine the distances or the sizes of objects as massive and as distant as the sun and the moon, and therefore, we do not see the size differences between the sun and the moon. For another example, the actor Peter Dinklage standing close to you might make the same visual angle as the retired basketball player Shaquille O’Neal would standing some distance from you. Similar visual angles from objects of different sizes and differ- ent distances is also the point of Figure 7.33. The point of Holway and Boring’s (1941) experiment was that depth provides information that allows us to appropriately scale the size of an object. Go to ISLE 7.11 for a demonstration of the relation of visual angles and size perception. TEST YOUR KNOWLEDGE 1. What is size–distance invariance? How is it used by the visual system to determine size? 2. Diagram what happens to the visual angle of an object as it gets farther from you. SIZE CONSTANCY 7.6 Diagram the concept of size constancy and how it functions in our visual system. Visual angle Size of retinal image Size of retinal image (a) (b) Visual angle FIGURE 7.32 Visual Angle and Size Visual angle is a function of the size of an object and its distance from the observer. When the object moves closer, its visual angle on the retina increases. When the object moves more distant, its visual angle decreases. If we know the object, we will see these differences not as changes in size but as changes in distance. Visual angle: the angle of an object relative to the observer’s eye ISLE 7.11 Retinal Image Size and Distance 213 Chapter 7: Depth and Size Perception Size constancy is the perception of an object as having a fixed size, despite the change in the size of the visual angle that accompanies changes in distance. That is, we have a tendency to see an object as the same size regardless of the size of its image on our retinae. As we have seen, there are limits to size constancy. Shaquille O’Neal does not look so tall when we see him from far away. However, at the range of normal viewing, size constancy allows us to see objects as the “right” size even as they move away from or toward us (Figure 7.34). For example, consider talking to a friend. You see her as the same size if she is standing 6 feet away from you or if she is standing 3 feet away from you. If she is standing 6 feet away and suddenly takes a step closer to you, you do not see her grow, even though her visual angle is now twice as large on your retina. Thus, size constancy usually allows us to see objects as not changing in size, but there are also exceptions to it. Consider the boot in Figure 7.7c. If we saw this boot in isolation, we might think it was an ordinary boot. Only in the context of all the cues that tell us that it is an enormous boot do we see it as such. We now turn to a set of illusions that occur when depth and size information are not well specified in a visual image. TEST YOUR KNOWLEDGE 1. Describe size constancy and why it is important for perception. 2. Under what circumstances might size constancy result in illusions rather than the perception of reality? VISUAL ILLUSIONS OF SIZE AND DEPTH 7.7 Demonstrate your understanding of illusions of size and depth and how they help us understand the operation of our visual system. Visual illusions fascinate us because they cause us to see that which is not really there. We see colors when no colors exist, we see spots of light where no such spots exist, and we FIGURE 7.33 Size and Visual Angle Illusions (a) Is the author (JHK) really a giant and as tall at the pyramid in front of the Louvre? The lack of cues for distance suggest that he’s as tall as the pyramid and standing right next to it. (b) The effect is similar in the picture of the woman “kissing” the sphinx. She is really much closer to the camera than to the sphinx, but the lack of depth cues makes it appear that she is even with the sphinx and big enough to kiss it. M ar ga re t K ra nt z RE U TE RS /J am al S ai di FIGURE 7.34 Size Constancy at the Range of Normal Viewing Shaquille O’Neal talks to the crowd during the 2015 March Madness Music Festival, April 5, 2015, in Indianapolis, Indiana. Scott Legato/W ireIm age/G etty Im ages Size constancy: the perception of an object as having a fixed size, despite the change in the size of the visual angle that accompanies changes in distance (a) (b) 214 Sensation and Perception see shapes that cannot be. Illusions of size and depth are no exception. By manipulating the depth cues we have discussed in this chapter, it is possible to get viewers to see depth relations that are really not present in an image. The Müller–Lyer Illusion In the Müller–Lyer illusion (Figure 7.35), we see the left line as longer than the right line, even though both lines are exactly the same length. If you do not believe this assertion, measure the two lines with a ruler. They are the same length. We will not distort illusions in this textbook, but please do not trust us. Measure our illu- sions objectively, and you will discover how your perception can be tricked. With respect to the Müller–Lyer illusion, the smaller lines that split off from the main line create the illusion that the main line is longer or shorter. An obvious question is why would these additional lines affect our perception of the length of the longest line? Most explanations of the Müller–Lyer illusion focus on the relation of size and depth. For example, Gregory (1966) advanced the view that the Müller–Lyer illusion is the result of misapplied size constancy. What this means is that the visual system wants to keep objects of the same size looking the same size, but in the case of the Müller–Lyer illusion, we mistakenly see size differences when the size is actually the same. The argument works as follows. Consider the left image in the Müller–Lyer illu- sion. Think about what this might look like in three dimensions. We may see the Müller–Lyer image as a corner in a wall. If the corner comes toward us, then the little projections at the top and bottom will move toward the middle of the line similar to receding parallel lines in linear perspective, as the walls lead away (see the left-hand image of Figure 7.36). In the image on the right of Figure 7.36, we see the corner pointed away from us, and the little projections that mark the corner are coming toward us. To draw them, they will go away from the line. Here we see the corner as being at a distance and the walls coming toward us. Because we see the line as being farther away in the right-hand image than we do in the left-hand image, we see the line in the right-hand image as longer. Why? Because it takes up the same space as the left-hand line on our retina, but we perceive it as being more distant. Objects that take up the same amount of space on our retina, but are more distant, are necessarily larger. Hence, we see the line as longer. You can see how this looks in real life by examining the corners in Figure 7.36. However, the previously stated explanation stemming from Gregory’s (1966) work is not the only explanation for the Müller–Lyer illusion. Some researchers have questioned whether an implicit assumption of depth is required to see the Müller–Lyer illusion. Day (1990) proposed a simpler explanation of the Müller– Lyer illusion. Day claimed that because the extending lines reach up and out in the right-hand image of Figure 7.36, the overall figure is longer in that image than the one on the left-hand side. Thus, we perceive components as longer as well. This view accounts for the Müller–Lyer illusion without recourse to a depth illusion. Another interesting observation about the Müller–Lyer illusion that suggests that Gregory’s account might not be complete is that you still get the illusion when done by active touch (haptically). You can demonstrate the Müller–Lyer illusion in both blindfolded and even congenitally blind participants. You build the Müller– Lyer stimulus using rods, with a central rod whose length is to be estimated and rods at the end forming the wings like in Figure 7.36. Now you have them esti- mate the length of the central rod by moving their hands along the rod stimulus FIGURE 7.35 The Müller–Lyer Illusion We see the vertical line as longer in the right figure than the left figure even though the vertical lines are the same length in both. Müller–Lyer illusion: the illusion where a line that has two lines going away at an angle looks longer than a line of the same length but the end lines angle back across the main line Ponzo illusion: the illusion in which two horizontal lines are drawn one above the other; both lines are on top of two inwardly angled vertical lines; the top line, where the two vertical lines are closer together, looks longer 215 Chapter 7: Depth and Size Perception and people still have the illusion. There are no depth cues here (Heller et al., 2002). Go to ISLE 7.12 for interesting variants on the Müller–Lyer illusion. The Ponzo Illusion The Ponzo illusion is a strong example of misapplied size constancy via the influence of linear perspective (a monocular cue to depth) on size perception. The Ponzo illusion is illus- trated in Figure 7.37. The two cows are the identical size (indeed, the identical animal). They take up exactly the same amount of size on the page or screen. If you do not believe this, you can measure them. However, the cow closer to the top of the page looks bigger. This is because there are a number of cues in the photograph that give clear indications of depth relations. We see the roughly parallel lines of the side of the road receding into the distance. We see a texture gradient of flowers and grasses. We see that the trees at the top of the photo are a bit hazy. Familiar size cues suggest that the trees, despite their smaller retinal image than the grass, must be larger but more distant. As a function of all this, we see the upper cow as farther away from the viewer than the lower cow. Because the cows take up the same amount of space on the retina, and the upper cow is farther away, the visual system makes the inference that the upper cow must be larger, leaving us with the strong perception of a big cow and tiny cow in front of her. The Ponzo illusion is so called because it was discovered by the Italian psychologist Mario Ponzo (1882–1960). For more on the Ponzo illusion, go to ISLE 7.13. The Ames Room Illusion The Ames room is a neat illusion because it can be instan- tiated in real space, as has been done in numerous sci- ence museums, such as those in Melbourne, Australia; Keswick, England; and Jerusalem, Israel. It also works on paper (Figures 7.38 and 7.39). In the Ames room, we put two people of normal size into a room that is anything but normal. When viewed through a peephole on the side of one of the walls, the perception is of a very large person in one corner and a very small person in the other corner. The room, however, looks normal. But the room is not normal. Indeed, it defies almost all conventional rules of rooms. Its walls (except the wall people are look- ing through) are trapezoids, as are the windows in the room. In versions in which there are floor tiles in the room, these are also trapezoids (see Figure 7.38). However, when we FIGURE 7.36 The Müller–Lyer Illusion in the Real World Note in the first picture that the angled lines indicate an outside corner, whereas the angled lines in the second picture indicate an inside corner. The inside corner with the walls coming toward us suggests that the straight line is farther away from us, and thus we perceive it as bigger. ISLE 7.12 Müller–Lyer Illusion FIGURE 7.37 The Ponzo Illusion Even though both cows take up the exact same space on the retina, the one higher on the image looks bigger because it appears farther away. ISLE 7.13 Ponzo Illusion 216 Sensation and Perception look through the viewing hole, we cannot distinguish the trapezoids, and we use our cues of familiar sizes and shapes to infer a normal room. Our visual systems convert the trapezoids into squares, so that we can perceive the room itself as normal. Because we perceive a normal room, we must therefore infer that the people are abnormally short or tall. To make the illusion fun, you can find twins who are dressed alike and put one at each end of the room, making for a very tall twin and a very short twin. You can see a video of people moving through the Ames room in ISLE 7.14. The Moon Illusion The moon illusion is another powerful illusion related to size–depth relations. What is interesting about this illusion is that we can see it in the night sky several times each month, which makes it a rare illusion in that it happens in nature. Find out when the next full moon is, or a day before or after the full moon. Then find out when sunset will occur, and watch the moon rise at sunset and then look at it again a few hours later. When we see the full moon on the horizon (such as when it rises at sunset and sets at sunrise), we perceive it as being larger than when it is higher up in the sky. This perceptual fact is seen in countless romantic movies, in which a love-struck couple stares at an abnormally large full moon. The moon illusion is depicted in Figure 7.40. As you can see, in Figure 7.40, the moon looks bigger at the horizon than at zenith. However, the moon does not change size across the night. Its size remains constant, and its distance from the Earth remains constant. If you do not believe this, take a photograph of the full moon as it rises and then again a few hours later. Print out the photograph and cut out the image of the moon. You will see that the zenith moon fits perfectly on top of the hori- zon moon. The moon does not change size; only our perception of it does. FIGURE 7.38 The Ames Room Illusion The people in the photo are the same size, but the person on the left appears much larger than the other because of the shape of the room. Note that the person on the left in one photograph is the person on the right in the other photograph. Did she shrink while her friend grew? ST EP H A N IE P IL IC K/ St af f/ G et ty Im ag es Ames room: a specially constructed room where two people of the same size standing in the two back corners will look very different in height Moon illusion: the illusion where the moon looks larger when it is near the horizon than it does when overhead Peephole Twice as far from observer as the man on the left FIGURE 7.39 The Ames Room Revealed You can see in this drawing what an Ames room looks like. When you look through the peephole, as the observer is doing, it looks like Figure 7.38, but you can see here that the person on the right is approximately twice as far away from the observer as the person on the left. But for the observer, this difference is not seen, and instead the person on the left is seen as much bigger. ISLE 7.14 Ames Room ISLE 7.15 Moon Illusion 217 Chapter 7: Depth and Size Perception FIGURE 7.40 Moon Illusion You can see here that the moon looks bigger when we see it on the horizon (a) than when it is higher in the sky (b). But if you match the moons, you will see that they are the same size. © iS to ck ph ot o. co m /k ei ic hi hi ki (a) © iS to ck ph ot o. co m /k la gy iv ik (b) The explanation for the moon illusion is similar to other size–depth illusions: a misperception of distance causes a change in the perception of size. We see the sky as a giant dome overhead, and objects in the sky as all at the same distance, that is, “painted” on the dome (though intellectually we know this to be untrue). As such, on this dome, the horizon is farther away from you than is the zenith, directly overhead. Because we perceive the horizon as being farther away than the zenith overhead, an object that takes up the same amount of space on our retina must be larger. At zenith, the moon looks smaller because the same-size object is now thought to be closer. This explanation is illustrated in Figure 7.41. You can explore an interactive illustration of the moon illusion on ISLE 7.15. TEST YOUR KNOWLEDGE 1. Assess the role of depth perception in the Müller–Lyer, Ponzo, Ames room, and moon illusions. 2. What is the Ames room? How does it illustrate the cue approach to depth and size perception? FIGURE 7.41 The Moon Illusion Explained We view the sky as a dome, and the moon and stars as “painted” on the dome, even though we know intellectually that this is not true. Thus, because the horizon is perceived as farther away than the zenith, the same-size object is viewed as larger on the horizon than at the zenith. Actual path of moon Horizon (not drawn to scale) Apparent path of moon Apparent size and distance of moon directly overhead Apparent size and distance of moon near horizon Same visual angle 218 Sensation and Perception EXPLORATION: Stereopsis and Sports: Do We Need Binocular Vision in Sports? In everyday life, we can get by with monocular cues, but the binocular cues add to our perception of depth and give us additional invaluable cues, which in some cases may help us perceive depth. But are there situations in which having binocular vision is absolutely essential, and are there situations in which binocular vision gets in the way? Sports provide one of these situations—competing in sports requires finely honed senses and close coordina- tion between what one is seeing and what one is doing. Think of the coordination between quarterback and wide receiver in football. The quarterback must time his pass at just the right time to avoid the hands of oncoming defend- ers and in such a way as to bring the football into contact with the wide receiver’s hands at just the right moment. Similarly, the wide receiver must be aware of the positions of defenders, the pattern he is supposed to run, and the exact arrival of the football at a particular point in time. Moreover, the wide receiver must do this while running nearly 20 mph (32 km/h). One would think that such activities require excellent depth perception (and periph- eral vision), and yet the former football player Wesley Walker succeeded at being a wide receiver despite being legally blind in one eye. However, Walker was a runner as fast as Olympic sprinters, which gave him an advantage despite his monocular vision. However, in target sports, the conventional wisdom is that it is desirable to eliminate binocular cues. Archers and tar- get shooters are usually instructed to close one eye. The thinking here is that good shooting means focusing on lining up a close point, such as your gun’s sight, and a far point, the target. Because the close point may be on noncorresponding points on your left and right retinae, it may not be possible to line up the sight using both eyes, especially if firing a handgun. Given that shooters in com- petitions are usually aiming at targets that are largely out of the range of binocular depth cues anyway, there is no real advantage to using both eyes, but there is this one big disadvantage. Unfortunately, there has been no published empirical research on the advantages or disadvantages of stereopsis in target sports. But are there any empirical data athletes could rely on? Do we know that quarterbacks, for example, use binoc- ular cues of depth to judge distance, and do we know if archers would shoot better if they kept both eyes open? In this section, we discuss some of the existing literature on stereopsis and sports. Much of the research has been done with respect to the batter in baseball. Consider the perceptual-motor task a baseball batter is confronted with. A pitcher is throwing a ball with a 9-inch circumference (230 mm) at more than 90 mph (145 km/h) from a distance of 60 feet (18 m). Baseball hitters must react extremely quickly; indeed, they often initiate their swing even before the pitcher has released the ball. As the ball approaches the batter, there are a number of depth cues that the batter may attend to, including familiar and relative size, linear perspective, shad- ing, optic flow, and binocular disparity. So, our question is this: To what extent do baseball batters use binocular disparity information? In one study, Hofeldt and Hoefle (1993) compared profes- sional baseball players in the highest professional league (the major leagues) with professional baseball players who had not made the highest professional league (the minor leagues) (Figure 7.42). Hofeldt and Hoefle looked at the players’ ability to do a stereo matching task that measured individual differences in the ability to use stereoscopic cues. The researchers found that the major league play- ers were more accurate at the stereoscopic task than the minor league players and that ability in the task predicted batting average (a measure of baseball performance) in the major leaguers. This suggests that baseball batters do use stereo cues to time their swings. However, this study is still correlational in nature. We do not know if batters are better because they have better stereopsis, or if better stereopsis is the result of working harder to be good at baseball. So Hofeldt and his colleagues next designed an experimental procedure to look at baseball batting and depth perception. Hofeldt, Hoefle, and Bonafede (1996) examined ama- teur baseball players hitting in batting cages under con- trolled conditions, including a constant speed of pitches. Again, they were interested in the interaction of the two eyes in timing the swing of a baseball bat. Baseball batting success was measured by looking at the rate of hits (contact with the ball into play), fouls (contact with 219 Chapter 7: Depth and Size Perception the ball, but out of play), and misses (not making con- tact with the ball). They then used filters to impair one eye or the other during swinging. These filters reduced the amount of light reaching the filtered eye. When one eye receives less light than the other, that filtered eye essentially perceives more slowly, making it harder to match corresponding points. This effect can also change the apparent depth of object in an illusion called the Pulfrich effect. The researchers found that when both eyes were filtered, the decrease in hitting success was negligible and nonsignificant. However, when one eye was filtered, but not the other, motion-for-depth cues were interrupted, and hitting success decreased dramat- ically, from 87% hits to 36% hits. However, this was when the dominant eye was filtered. The effect was smaller (80%) when the nondominant eye was filtered. Scott, van der Kamp, Savelsbergh, Oudejans, and Davids (2004) also found that participants were better able to hit a foam object thrown at them if they had access to binocular cues than when they only had monocular cues. Hofeldt et al. concluded that (a) binocular information is important for baseball batting and (b) baseball batters really do have a dominant eye, consistent with the con- ventional sports wisdom. In baseball, players must be able to catch balls as well as hit them. Outfielders must judge the distance and direction a ball is going to go after the batter hits it, and they must also time their running speed to meet the ball in an antici- pated location. We can ask to what extent baseball players use stereopsis to judge where they must run to intercept the ball. Mazyn, Lenoir, Montagne, and Savelsbergh (2004) looked at the ability of people to catch a tennis ball with one hand. They compared performance under binocular viewing conditions as well as monocular viewing conditions at a variety of ball speeds. People with normal stereo vision caught more balls under binocular conditions than under monocular conditions, and this effect was more pronounced the faster the ball was thrown at them. Thus, binocular cues are used in catching balls, contrary to some earlier notions of how baseball players intercept fly balls (see Mazyn et al., 2004). Mazyn et al. also looked at people who do not see well in stereo. These individuals were no better in the bin- ocular condition than the monocular condition and were somewhat lower in their overall ability to catch tennis balls. Along similar lines, Isaacs (1981) found that people with better stereo vision were also better at shooting foul shots in basketball than people with weaker stereo vision. Thus, across an assortment of athletic tasks, we see that stereo vision is useful in performing athletic tasks. APPLICATION: Virtual Reality and Therapy Although not all that common yet, views like Figure 7.43a are becoming increasingly common. People are sticking their phones into strange head pieces, putting these head pieces on, and craning their heads all around. These head pieces or goggles use phones to create what is called virtual reality (see Figure 7.43b for an anaglyph of a couple of examples of different goggles that can be used). Virtual reality refers to a computer-generated three-dimensional photograph, image, or environment that can be interacted with in an apparently authentic way. To create convincing virtual reality, all the depth and size cues that we have discussed in this chapter have to be understood and reproduced accurately. For example, panoramic photographs are becoming quite popular, but what if you could put yourself back inside the photograph and turn around to see the different FIGURE 7.42 Binocular Vision in Baseball A batter must be able to accurately time an incoming baseball in a fraction of a second in order to hit it. Batters hit better with binocular vision than with monocular vision. © Shutterstock.com /sirtravelalot Virtual reality: a computer-generated photograph, image, or environment that can be interacted with in an apparently real way 220 Sensation and Perception directions in the original location? With virtual reality goggles, that is possible. Phones will draw two versions of the image and the headsets use the same technique as in Brewster stereoscopes to help you fuse the images so you can have the full range of monocular and bin- ocular depth cues. You can try some in ISLE 7.16. If you are on a laptop you can scroll the images around with your mouse. If you have a tablet, make the image full screen and as you turn around you will see dif- ferent views of the photographs. If you have gog- gles that can run Google Cardboard images, you can fully put yourself into the image which will be pre- sented stereoscopically. You look around by turning your head and body as if you were there. In a fully computer- generated version of virtual reality, the programmer must implement cues such as relative size. That is, if all of the people in a scene are drawn the same size, instead of smaller for more distant figures, then our size con- stancy will make those figures in the background loom as giants like this version of the Ponzo illusion (Figure 7.44), which might be interesting but violates what real- ity looks like. Virtual reality goggles are making commonly available what has been discussed in research for a long time in terms of the principles dis- cussed in this chapter. Virtual reality is also now available to anyone with a smartphone, and the price for the head- sets is dropping fast. As such, new ways to experience locations and play games will continue to emerge as the technology matures. Think of revisiting your vacation by putting on some goggles and being able to look around a scene just like when you were there. It appears that these goggles might be useful for more than reliving experi- ences or playing games. One interesting development is the use of virtual reality for psychological therapies. One area of therapy in which virtual reality is being explored as a tool is relief from anxiety and phobic (fear) disorders. All of us have anxiety from time to time. However, in people who have an anxiety disorder, anxi- ety persists and can interfere with day-to-day functioning. An anxiety disorder refers to a broad class of disorders in which anxiety persists and is greater than the situation seems to warrant. In a phobic disorder, the person has a fear of a situation or object. A common method of treat- ment, under various names and guises, can be called expo- sure therapy. In brief, the client is exposed in a controlled manner to the stimulus that is causing the anxiety or fear. The exposure starts at a level that the patient finds easy to handle and then as each level of exposure is handled with- out any increase in anxiety or fear, a more intense level of the phobic stimulus is presented. Virtual reality creates the possibility of controlled exposure to the anxiety or fear-inducing stimulus FIGURE 7.43 Virtual Reality Goggles (a) A photograph of one of the authors wearing virtual reality goggles and looking around. A smartphone provides the stimuli. (b) An anaglyph of a couple of different types of virtual reality goggles. Use your glasses to see them in depth. (a) (b) ISLE 7.16 Virtual Reality Photographs 221 Chapter 7: Depth and Size Perception through computer generation. When virtual reality is used, the therapy is called virtual reality exposure therapy (VRET). The idea here is that in a computer environment, the exposure can be much more carefully controlled. One anxiety disorder is post-traumatic stress disor- der (PTSD) in which people experience symptoms of a trauma long after an experience occurred. PTSD occurs not only for many battle-tested soldiers but also for civil- ians who have suffered trauma as a result of having expe- rienced assault, rape, or a natural disaster. Although still in pilot stages of testing, these virtual reality therapies have shown promise. In one study, Portuguese veterans were placed in a jungle setting where triggers for their PTSD were presented, such as a mortar explosion. Over repetitions, the strength of the trigger could be increased very precisely, thereby controlling the exposure (Gamito et al., 2010). In another study, elderly American veterans of the Vietnam war were flown over a virtual Vietnam and landed in a clearing in the jungle (Rothbaum et al., 1999). In both cases, the clients showed signifi- cant improvements, and Rothbaum and colleagues found that the improved symptoms lasted at least 6 months. However, the number of participants in both cases is small and the results are considered preliminary. You can learn more about the use of virtual reality therapy for PTSD in ISLE 7.17. Specific phobic disorders, such as arachnophobia (intense fear of spiders), is another area where the use of virtual reality has shown promise. In traditional therapies, it often requires starting with something that distantly suggests the phobic stimulus and gradually getting ever closer to the dreaded arachnid. Several studies have recently exam- ined using VRET for specific phobias, including arach- nophobia (e.g., Bouchard, Côté, St-Jacques, Robillard, & Renaud, 2006). There are a sufficient number of studies on this topic that it is possible to see if different researchers are finding consistent results (Opriş et al., 2012). It seems that these virtual reality therapies are at least as effective as the best traditional therapies on several measures, such as reduction of symptoms and dropout rate. The question then arises, why do the virtual reality therapy? Bouchard and colleagues (2006) might have an answer to this question. Developing virtual reality programs is quite expensive. However, so, too, are traditional therapies. Overcoming a phobia such as arachnophobia might even- tually require having to deal with spiders. Virtual spiders are easier to maintain and don’t require lots of mosquitos. Bouchard used an existing virtual reality game to help in the treatment of arachnophobia. Using an existing game can reduce cost, and if these emerging goggles continue to reduce in cost as well, this form of therapy might become cost-effective. You can see more about the use of virtual reality in treating phobic disorders in ISLE 7.17. All in all, it appears that these funny-looking goggles might be more than toys. FIGURE 7.44 Watching Out for Relative Size In virtual reality, all of the depth cues in the real world have to be accurately reproduced. For example, if you do not make more distant figures smaller, they become giants as in this version of the Ponzo illusion. ISLE 7.17 Virtual Reality and Therapy Sensation and Perception222 CHAPTER SUMMARY 7.1 Explain monocular depth cues and how they work to allow us to perceive depth with one eye. The problem our visual systems must solve is that we rep- resent the world on flat, 2D retinae. However, we live in and therefore must perceive a 3D world. Our visual system seems to do this in classic Helmholtzian fashion—by using unconscious inference from myriad clues that inform us of relative depth. We divide these cues into three groups— the monocular cues, the oculomotor cues, and the binoc- ular cues. The monocular cues include information about depth that we can infer with just one eye. These tend to be characteristics of the scene itself or motion-based cues. Monocular cues include occlusion, relative height, relative size, familiar size, texture gradients, linear perspective, atmospheric perspective, shading, and shadows. Included in the monocular cues are the motion-based cues of motion parallax, accretion and deletion, and optic flow. 7.2 Compare the different oculomotor curves in their effectiveness as depth cues. The oculomotor cues are convergence (or vergence) and accommodation. Convergence means that the eyes rotate toward each other when looking at closer objects. We sense this motion, and it informs us about whether we are looking at a near or a far object. Accommodation is the process of adjusting the lens of the eye so that both near and far objects can be seen clearly. Again, we can sense the muscles that make this adjustment, and this informs us about the proximity of the objects that we are looking at. 7.3 Summarize the principle of stereopsis and how it applies to human depth perception. Stereopsis is the strong sense of depth we perceive from the visual system’s processing of the comparison of the two different images from each retina. Stereopsis arises from binocular disparity. Binocular disparity occurs because our two eyes are in different locations in our heads and there- fore have slightly different views of the world. For stereop- sis to work, the visual system must match the image on the left retina to that on the right retina. 7.4 Sketch the correspondence problem and how it must be solved for stereopsis. Corresponding points refers to a situation in which a point on the left retina and a point on the right retina would coin- cide if the two retinae were superimposed. In contrast, noncorresponding points refers to a situation in which a point on the left retina and a point on the right retina would not coincide if the two retinae were superimposed. Corresponding points lie along the horopter, the region in space where the two images from an object fall on corre- sponding locations on the two retinae. We use disparity information to determine if an image is in front of the horop- ter or behind it. Crossed disparity means that an object is in front of the horopter, whereas uncrossed disparity means that an object is behind the horopter. The correspon- dence problem is the problem of determining which image in one eye matches the correct image in the other eye. Stereograms and anaglyphs are specially designed pictures in which the photograph or film replicates the binocular dis- parity and therefore creates a pop-out phenomenon that we see as depth. Random-dot stereograms have been used to show that object recognition is not necessary for disparity cues, thus placing the solution to the correspondence prob- lem as something that occurs prior to object recognition. Physiological studies demonstrate that there are binocular cells in the visual cortex. These cells are tuned to disparity information, thus allowing the neurons to extract informa- tion about depth. Infants do not develop stereopsis until the age of about 4 months, though the physiological reasons for this are unclear. 7.5 Explain the concept of size perception and the inferential nature of its determination. Size and depth are deeply intertwined. Size–distance invariance refers to the relation between perceived size and perceived distance, and simply states that the per- ceived size of an object depends on its perceived distance, and the perceived distance of an object may depend on its perceived size. 7.6 Diagram the concept of size constancy and how it functions in our visual system. Size constancy is the perception of an object as having a fixed size, despite the change in the size of the visual angle that accompanies changes in distance. 7.7 Demonstrate your understanding of illusions of size and depth and how they help us understand the operation of our visual system. A number of illusions demonstrate violations of size–distance invariance and size constancy. In the Müller–Lyer illusion, we see a line that looks longer than an identical-length line Chapter 7: Depth and Size Perception 223 because the ends of one line suggest a corner moving away in depth. The Ponzo illusion is a strong example of misap- plied size constancy as well as the influence of linear per- spective (a monocular cue to depth) on size perception. In the Ponzo illusion, we see two objects that take up the same amount of space on our retinae as different in size because of their perceived differences in depth. In the Ames room, we put two people of normal size into a room that is anything but normal. When viewed through a peephole on the side of one of the walls, the perception is of a very large person in one corner and a very small person in the other corner. The room, however, looks normal. The moon illusion is another powerful illusion related to size–depth relations. When we see the full moon on the horizon (such as when it rises at sunset and sets at sunrise), we perceive it as being larger than when it is higher up in the sky. REVIEW QUESTIONS 1. How do jumping spiders use their multilayered ret- inae to perceive depth? How is this different from mammalian depth perception? 2. What are the oculomotor cues to depth? How do they work, and how do they give us depth information? 3. Describe three pictorial monocular depth cues. How does each one provide us information about depth or distance? 4. What is motion parallax? How is it used as a cue for depth? 5. What is meant by the term stereopsis? Why does stereopsis require two frontally placed eyes? 6. What is the horopter? How do crossed and uncrossed disparity relate to the horopter? What is meant by zero disparity? Where would such a point lie along the horopter? 7. What is the correspondence problem? How do ran- dom-dot stereograms address the theoretical issues raised by the correspondence problem? 8. What evidence suggests that stereopsis is not innate? What areas of the brain appear to be involved in stereopsis? 9. What is meant by size–distance invariance? What is size constancy? Describe one illusion that illustrates each principle. 10. What data exist to support the idea that stereop- sis can help athletic performance? Choose a sport not considered in this chapter (e.g., volleyball) and describe how depth perception might be critical to perform well in that sport. PONDER FURTHER 1. In many ways artists are our first developers of virtual reality. On flat canvases and other surfaces, they cre- ate the impression of depth, and yet they are restricted to monocular depth cues. Survey artwork online, or by any other way you have access to it, and determine the depth cues that artists use to create the percep- tion of depth. 2. Review ISLE 7.8, all parts but particularly ISLE 7.8b. Panum’s area of fusion suggests that we get our best depth perception from binocular disparity in a very nar- row range around the horopter. Large differences in depth do not benefit nearly as much from our having binocular disparity. So, consider 3D movies that use binocular disparity. You are giving advice to the director on the most effective use of binocular disparity for a 3D movie. What would you tell this director? KEY TERMS Accretion, 197 Ames room, 215 Atmospheric perspective, 194 Binocular cells, 209 Binocular disparity, 201 Correspondence problem (depth perception), 206 Corresponding points, 202 Crossed disparity, 203 Sensation and Perception224 Cue approach to depth perception, 189 Deletion, 197 Diplopia, 203 Familiar size, 192 Horopter, 202 Linear perspective, 193 Monocular depth cues, 189 Moon illusion, 216 Motion parallax, 196 Movement-based cues, 189 Müller–Lyer illusion, 214 Noncorresponding points, 202 Occlusion, 190 Optic flow, 198 Panum’s area of fusion, 203 Pictorial cues, 189 Ponzo illusion, 215 Random-dot stereograms, 208 Relative height, 191 Relative size, 191 Shadows, 195 Size constancy, 213 Size–distance invariance, 211 Stereopsis, 200 Texture gradient, 194 Uncrossed disparity, 204 Vergence, 199 Virtual reality, 219 Visual angle, 212 Zero disparity, 204 Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review, study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning. Learning Objectives Digital Resources 7.1 Explain monocular depth cues and how they work to allow us to perceive depth with one eye. 7.2 Compare the different oculomotor curves in their effectiveness as depth cues. Extremal Edge: A Powerful Cue to Depth Perception and Figure-Ground Organization 7.3 Summarize the principle of stereopsis and how it applies to human depth perception. Seeing in 3-D With Just One Eye: Stereopsis Without Binocular Vision Going Binocular: Susan’s First Snowfall Losing Stereo Vision and Depth Perception—a Personal Account From Oliver Sacks Jumping Spiders 7.4 Sketch the correspondence problem and how it must be solved for stereopsis. 7.5 Explain the concept of size perception and the inferential nature of its determination. Distance Perception 7.6 Diagram the concept of size constancy and how it functions in our visual system. 7.7 Demonstrate your understanding of illusions of size and depth and how they help us understand the operation of our visual system. Dynamic Müller Lyer Illusion The Dynamic Ebbinghaus Tim Flach/Iconica/Getty Images 8Movement and Action Mads Perch/Stone/Getty Images LEARNING OBJECTIVES 8.1 Explain the concept of a motion threshold and what variables influence that threshold. 8.2 Describe the anatomy of the visual system dedicated to motion perception. 8.3 Discuss how we infer form from the perception of biological motion. 8.4 Summarize the concept of an affordance. INTRODUCTION A number of major world cities, such as Los Angeles, San Francisco, Port-au-Prince (Haiti), Tokyo (Japan), Reykjavik (Iceland), Ankara (Turkey), and Wellington (New Zealand), to name just a few, lie in earthquake zones. Anyone who has ever lived through an earthquake understands how motion can induce fear. Once the earthquake starts, things that are not supposed to move, like walls and floors, may start shaking, undulating, and cracking. Other things start falling—books, jars, glasses, and televi- sions. People outside their homes may see bridges sway and roads crumble. Cities heav- ily fortified against earthquakes (e.g., San Francisco, Tokyo) design their buildings to sway in an earthquake, so as to avoid cracking and crumbling. However unnerving this may be for residents of these cities, it saves countless lives when these highly popu- lated cities are struck by earthquakes. But simply watching swaying buildings is rather unnerving. Only when all this motion subsides can people relax. What many find strik- ing about perceiving motion during an earthquake is that objects we generally perceive as stable—walls, roads, and buildings—move. This type of motion induces fear. Clearly, perceiving motion is important and vital for human beings. Imagine now that you are on an airplane, perhaps one bound from San Francisco to Tokyo on your way to a conference on sensation and perception research. You are 7 miles above the Pacific Ocean, traveling at nearly 1,000 km/h (621 mph). A seren- dipitous error on the airline’s part bumped you up into first class. Therefore, you are sitting in a comfortable reclining chair, eating warmed nuts and drinking champagne, completely unaware that you are moving at nearly the speed of sound (which is 1,236 km/h). Why is it that you are not perceiving the incredible motion that is occurring in this situation? With the shades drawn shut, you do not see your motion relative to the ground below. You see only the tranquil scene of people relaxing in their first-class seats, feeling smug that they are comfortable while the rabble behind them is not. Your world is stable and not moving. ISLE EXERCISES 8.1 Relative Motion 8.2 Time-Lapse Motion 8.3 Motion Thresholds 8.4 Apparent Motion 8.5 Correspondence Problem in Motion (Wagon Wheel Effect) 8.6 Induced Motion 8.7 Reichardt Detectors 8.8 Corollary Discharge 8.9 Eye Movements 8.10 Correlated Motion 8.11 Motion Aftereffect 8.12 Structure From Motion 8.13 Biological Motion 8.14 Cat and Laser Pointer 8.15 Optic Flow 8.16 Rotating Snakes 8.17 Spiral Staircase 8.18 Illusory Rotation 8.19 The Furrow Illusion 228 Sensation and Perception You may remember from physics class that all motion is relative. In the case of the airline passenger, everything around her is moving at the same speed. The comfortable chair, the champagne glass, the magazines, the flight attendant, and the passenger are all moving together at the exact same speed. We may feel the rumbling of the engines and the rumbling in our stomachs when the plane encounters turbulence. But we feel this motion rather than see it. Unless you look out the window, you will not perceive motion visually under these circumstances. It is not all that dissimilar to why we do not perceive the rotation of the Earth itself (see ISLE 8.1). Everything around us, including us, is rotating with the Earth, so we do not perceive the motion. In this sense, we can define motion as a relative change in position over time. Our visual systems must be able to attend to motion. Perceiving motion allows us to move ourselves as well as to time the arrival of incoming objects, whether they are cars, baseballs, cheetahs, or falling coconuts. One can imagine that motion perception evolved to help our distant ancestors perceive incoming predators and fleeing prey. Nowadays, motion detection is important in driving cars, crossing city streets, playing sports, and watching movies. Thus, at its basic level, there are a number of features our visual system must be able to perceive. First, we need to know if moving objects are approaching us or heading away from us. In many cases, such as a fast-moving train, we need to know if an object is heading toward us so that we can get out of its way. In other cases, we need to know if an object is moving away from us, so that we can slow it down, as in the case of a child trying to catch an ice cream truck. Thus, the direction of motion, toward us or away from us, is critical. But we must be able to perceive this motion across three dimensions— directly ahead of us, to our sides, and above (or below) us. An object going straight up may not represent a danger, but an object coming down, and being accelerated by gravity as it does so, may represent a danger. Finally, we must also be able to estimate the speed of moving objects. A woman and child walking may not constitute a danger, but speeding race cars may very well be a danger (Figure 8.1). You can work with an interactive illus- tration of relative motion and frame of reference on ISLE 8.1a. HOW DO WE PERCEIVE MOTION? 8.1 Explain the concept of a motion threshold and what variables influence that threshold. Motion Thresholds: How Slow and How Fast? Think about watching a spinning bicycle wheel. When the rider first starts peddling, we see the motion of the spokes of the wheel. As the bicyclist pedals faster, the spokes start spinning faster and faster. We can see this motion, but eventually, the wheel is moving ISLE 8.1 Relative Motion Motion: a change in position over time © iS to ck ph ot o. co m /m ev an s (b) © iS to ck ph ot o. co m /m on ke yb us in es si m ag es (a) FIGURE 8.1 Motion Is Relevant (a) We see this woman and her child approaching us. Their speed, however, probably does not cause concern. (b) These race cars bearing down on us would cause concern and, hopefully, a sudden leap off the road. 229 Chapter 8: Movement and Action so fast that we can no longer track the individual spokes. Instead, we see a blur of motion. There are two points to be made here. First, some motion is so slow that we cannot perceive it, and other motion is so fast that we cannot perceive it. Think of a plant bending its leaves toward the direction of sun- light. Most plants move their leaves extremely slowly, so slowly in fact that human eyes do not perceive the motion. Only when we view a plant through time-lapse photography do we know that it is actually moving (Figure 8.2) (ISLE 8.2). Similarly, movements that are extremely fast, such as the spinning of an automobile’s wheel, may simply be too fast for us to perceive. Many sources of light flicker, such as our computer screens, but they do so at such a rapid rate that we do not perceive the change. Motion thresholds are a function of what parts of the retinae are seeing the motion. Interestingly, we have rather poor motion thresholds in the foveal regions of our ret- inae. But our motion thresholds are much lower (i.e., better) than our acuity in the periphery of our retinae. You may occasionally become aware of this, when you detect motion to your side and move your eyes and head to engage with the oncoming motion. Motion thresholds also nicely illustrate the importance of peripheral vision. Most of the functions of the visual system already discussed in this book have involved the region of the visual field near and around the fovea. Color and much of the discussion of form relate to the fovea and near it. Yet motion is a function that does very well at the periphery. It is tempting to think of our peripheral vision as merely poor vision, or a poor version of how vision operates in the fovea. However, that would be a mistake. Vision at the periphery is not so much poorer as it is different. The relative importance of acuity and motion in the fovea and at the periphery gives some insight into this difference. A moving stimulus at the periphery seems to grab the attention, and it may direct the fovea over to this stimulus. As with any other perceptual feature of vision, we can ask a few basic psychophys- ical questions about motion thresholds. For example, how slow can motion be if we are to still perceive it as motion? Do we see snails moving or plants shifting their leaves toward the sunlight? The answer may be yes for the snails but no for the plants. Because snails are small objects that move slowly, their motion has few consequences for us, so we may not ordinarily attend to snails in motion. But if we do, we can probably detect their motion as they slime their way up the sides of our houses. We know that for us to detect movement, an object must move at least 1 minute of 1 degree across the retina to be detected (that’s 1/60 of 1 degree of visual angle). Thus, our absolute threshold for motion detection will be a function of the speed of the moving object and its distance from us. At the other end of the spectrum, we can ask, at what high speed do we lose the ability to track motion? Can we follow the motion of a supersonic jet? Can we follow the individual rotations of a car’s wheels spinning when the car is moving at 70 mph? We know that such thresholds will be context dependent, so specific answers cannot be given. The brightness of an object, the size of the object, and the amount of time the object is visible will influence motion thresholds. In general, we can usually perceive fast motion, even if we cannot track it. Nonetheless, some objects move at such high speeds that we just cannot see them (e.g., a bullet shot at 1,500 feet per second). You can examine your own motion thresholds at different locations on your retina in ISLE 8.3. FIGURE 8.2 Blue Hyacinth Blooming © iStockphoto.com /nickp37 ISLE 8.2 Time-Lapse Motion ISLE 8.3 Motion Thresholds 230 Sensation and Perception Real and Apparent Motion Motion in the world is created by the con- tinual change in position of an object rela- tive to some frame of reference. That is, we watch the cat run across the kitchen tiles. We watch the leaves of the trees bend back and forth with the sky in the background. We watch ducks swim across the pond. This is real motion. However, human beings also perceive a number of forms of illusory motion, that is, situations in which we perceive motion when none actually occurs. One form of this illusory motion is apparent motion. Apparent motion is the appearance of real motion from a sequence of still images (Figure 8.3; also see ISLE 8.4 to explore apparent motion). Apparent motion occurs whenever stimuli separated by time and location are actually perceived as a single stimulus moving from one loca- tion to another. Apparent motion is the basis of our sense of motion in watching videography and animation, and it forms the basis for much of our entertainment in the form of television, movies, and com- puter games. When we watch Iron Man flying across the sky in the movies, what we are seeing is apparent motion. The pat- tern of lights on your screen creates an illu- sion that an object is moving. Apparent motion includes beta motion (optimal motion), in which an object is perceived as realistically moving on the basis of what is actually a series of stationary images being presented sequentially (Figure 8.4). Beta motion is the basis of motion seen in movies. It is similar to but different from phi motion, the basis of motion in the displays in Las Vegas. In beta motion, the perceived motion is indistinguish- able from real motion, while in phi motion, you perceive motion but can see that the ele- ments, like the individual bulbs in the Las Vegas displays, do not move. In both beta motion and phi motion, the images are turned on and off quickly to induce the perception of motion. You can see both types of motion in ISLE 8.4 by changing how fast the dots go on and off. Phi motion forms the basis of billboard displays, such as the famous Times Square signs that depict stock market information. You can work with an interactive illus- tration of apparent motion, both beta and phi, on ISLE 8.4. Indeed, all the motion FIGURE 8.3 Kinetoscopic Record of a Man Sneezing Filmstrip of a man sneezing for Thomas Edison’s Kinetoscope. Jo hn S pr in ge r C ol le ct io n/ Co rb is H is to ric al /G et ty Im ag es Real motion: motion in the world created by continual change in the position of an object relative to some frame of reference Apparent motion: the appearance of real motion from a sequence of still images ISLE 8.4 Apparent Motion Flash FlashDark FIGURE 8.4 Apparent Motion Illustrated (Beta Motion) When these two lights are flickered on and off at just the right rate, it will look like a single light is jumping from one location to the other. 231 Chapter 8: Movement and Action illustrations on ISLE make use of apparent motion. This is true for any movement one perceives on a television or computer screen. If you recall from Chapter 7, we discussed the correspondence problem with respect to depth perception. In that case, the correspondence problem pertains to how the visual system knows how to match objects perceived in the left eye with objects per- ceived in the right eye. We discussed how the visual system matches objects in the left and right eyes before object recog- nition, as indicated by successful depth perception in random-dot stereograms. In this chapter, we briefly describe the correspondence problem as it applies to motion perception. In motion perception, the correspon- dence problem refers to how the visual system knows if an object seen at Time 1 is the same object at Time 2. In most normal vision settings, we know that an object is the same because most objects do not go through rapid changes across time when they are in motion. Moreover, we may use eye movements to track that motion across time. But consider watching a flock of birds move across the sky. In this case, it may be difficult to track a par- ticular bird, but we do not see some birds flying out of existence and others coming into existence. We see a pattern of birds flying. This means that there must be a way in which we keep coherence in motion perception. The correspondence problem is illustrated in Figure 8.5. You can see an illustration of the issues involved in the correspondence problem on ISLE 8.5, which shows the wagon-wheel effect. In the wagon-wheel effect, we per- ceive the direction of motion in the opposite direction of the actual motion. Another motion-based illusion is induced motion. Induced motion means that one moving object may cause another object to look like it is moving. The classic example of induced motion is the movement of clouds at night, which may make it seem as if the moon is moving (Figure 8.6). The clouds moving in one direction induce a perception that the moon is moving in the opposite direction. Similarly, the movement of your car makes it look as if the world is rushing by you, when in fact it is you in the car who is rushing by the world, and the surroundings are relatively motionless. Induced motion is illusory because we misperceive what is moving relative to the other object. Your car is really moving relative to the surface of the road, but you perceive the road as moving relative to your stable position in the car (ISLE 8.6). How different is apparent motion from real motion? When we watch a movie, we suspend our knowledge that we are just looking at a screen and we feel as if we are actually watching speeding planes and talking people. How close is it to real motion? One question we can ask is whether there are different neurocognitive systems involved in the perception of real and apparent motion. That is, does apparent motion look like motion because it activates the same neural networks as real motion does? This question was put to the test by Larsen, Madsen, Lund, and Bundesen (2006). Larsen et al. (2006) compared the neural responses to real and apparent motion by ISLE 8.5 Correspondence Problem in Motion (Wagon Wheel Effect) ?? A B C (a) (b) (c) FIGURE 8.5 The Correspondence Problem How do we know that what we see at Time 1 corresponds to what we see at Time 2? The pictured square moves up and to the left. To detect the motion, the visual system must match points along the square at Time 1 with points along the square at Time 2. Correspondence problem (motion perception): how the visual system knows if an object seen at Time 1 is the same object at Time 2 Induced motion: an illusion whereby one moving object may cause another object to look as if it is moving ISLE 8.6 Induced Motion FIGURE 8.6 Cloud Movement Makes the Moon Appear to Move © iS tockphoto.com /kdshutterm an 232 Sensation and Perception recording brain activity with functional magnetic resonance imaging (fMRI) tech- nology. In the apparent motion condition, participants watched a standard apparent motion display of lights flashing back and forth such that it looked as if one light was moving from one location to the other. In the real motion condition, participants watched an actual light move from one location to the other. And in a control con- dition, participants also saw two lights flashing, but the lights did not appear to be moving. Larsen et al. then compared the areas of the brain that were active during each stimulus. The fMRI brain scans found that the areas of the primary visual cortex that were responsive to apparent motion were the same as those responsive to real motion. These areas within the visual cortex were active during both forms of motion but not during the control condition. Therefore, the region in the cortex was responding to movement, not to flashing lights. For this reason, we can be reasonably certain that the apparent motion participants view in experiments generalizes to motion perception of actual moving objects. TEST YOUR KNOWLEDGE 1. What is the correspondence problem as it applies to motion? How is it best solved? 2. What is the difference between phi movement and beta motion? THE NEUROSCIENCE OF VISION AND MOTION 8.2 Describe the anatomy of the visual system dedicated to motion perception. Motion Detection in the Retina The retinae play an important role in early motion detection. We know that some types of amacrine cells in the retina are sensitive to motion (Masland, 2012). Thus, the begin- nings of analyzing motion occur quite early in the human retinae. We also see, in the optic nerve that leaves the retina, the beginning of the M pathway, which codes for motion in V1 and along the dorsal stream of vision (Tapia & Breitmeyer, 2011). The P pathway also contributes to motion perception, as we see moving objects in color. Thus, both pathways at the level of the retina contribute to motion perception. Some studies have looked at motion perception in other animals. In a series of clas- sic papers, Barlow and his colleagues investigated motion perception in the retinae of frogs. Barlow (1953), for example, found cells in the retinae of frogs that responded to the transition from light to dark and then to the opposite transition. Later, Lettvin, Maturana, McCulloch, and Pitts (1968) called these cells bug detectors because they were particularly good at identifying the movements of flies, a primary food of frogs. The Complexity of Motion As we often repeat in this book, perceptual processes are complex, even if we do them effortlessly. The neuroscience of the perception of motion is no exception. First of all, our bodies are constantly in motion—we must be able to discriminate our own motion, both intended and unintended, from the motion of other people, animals, and objects. Consider an eye blink—every so often we mistake our own eye blinks for 233 Chapter 8: Movement and Action moving objects, but not very often. Usually, our visual systems take into account our own motions and do not project them onto the perception of other objects. Even when we are perfectly still and watching a moving object, such as a bird flying across the sky, our visual system has a complex task. Think of a simple case of motion—a single object moving in a straight line across the visual field. This might be a bird flying across the sky or a person walking down the street. As we mentioned earlier, we must be able to judge the direction, in each of the three dimensions, and the speed of this object. As we discussed in the opening chapter, we must also be able to judge the time to impact, that is, how long it will take before the moving object smashes into us. A bird flying fast directly at you is not usually a desirable state of affairs, whereas a bird flying past you is harmless. Thus, the two elements of motion detection—direction and speed—must be coded at the neural level. Now consider how this might be done in the visual system. We start off with the simplest case. Your eyes are not moving. They are looking straight ahead and not track- ing a moving object. The object is passing in front of you, moving from left to right in a straight line, without dipping or wobbling (Figure 8.7). The visual system must be able to correlate an object at one time (T1) in one position (P1) with the same object at another time (T2) in another position (P2). It must also be able to distinguish this motion from motion going in the opposite direction, that is, starting at P2 at T1 and reaching P1 at T2. And to reiterate, we also want to know the approximate speed and direction of the moving object. Now think about this in terms of a neural circuit. We need to have neurons that have receptive fields that include the points P1 and P2 in the visual world. But we need these neurons to respond temporally as well. That is, we want neurons that are sensitive to an object at P1 at T1 and then at P2 at T2, some specific later time point. We need other neurons that fire when objects are at P1 at T1 and at P2 at T2 at a range of differences in (T2 – T1), that is, different speeds. We also need neurons sensitive to movement in the other direction, that is, if an object moves from P2 to P1 at a particular speed. This allows us to track motion in the other direction. One possible model for motion detection in neurons is shown in Figure 8.8. The neurons in this model have been called Reichardt detectors after the German psychophysicist who first hypothesized them (Reichardt, 1969). If you inspect Figure 8.8 carefully, you will notice a few important features of Reichardt detectors. In the model, you can see motion-sensitive neurons (labeled M for motion). These M neurons are tuned to activity first in receptive field P1 and then in receptive field P2, if there is a specific delay between the activity in the two receptive fields. The delay specifies the speed to which this neuron is sensitive. If the delay is too short or too long, the M neuron will not make as large a response. Similarly, if the mov- ing object is going the other way, this neuron will not respond. Thus, this simple model can account for both speed (as measured by delay) and direction (as measured by which receptive field is activated first). Of course, there will be other motion-sensitive neurons that will respond to different delays and different directions of motion. Motion-sensitive neurons with longer delay times are sensitive to slower motion, whereas motion-sensi- tive neurons with shorter delay times are sensitive to faster motion. We find these cells in the magnocellular tract. You can work with an interactive illustration of Reichardt detectors on ISLE 8.7. Are there neurons like this in the human visual system? It turns out that there are many neurons in V1 that show these characteristics. V1 neurons, in addition to the features already discussed, seem to be tuned to specific directions and speeds of motion (Orban, Kennedy, & Butler, 1986; Wang & Yao, 2011). In nonprimate species, such as cats, there are also motion-sensitive neurons in the lateral geniculate nucleus (Wang & Yao, 2011). Moving spot of light Motion in retinal image d d FIGURE 8.7 Motion on the Retina A moving object (here moving from left to right) moves a certain distance across the visual field. The distance correlates with a distance across the retina. Reichardt detectors: neural circuits that enable the determination of direction and speed of motion by delaying input from one receptive field, to determine speed, to match the input of another receptive field, to determine direction ISLE 8.7 Reichardt Detectors 234 Sensation and Perception Corollary Discharge Theory Reichardt detectors work well when the eyes are stationary, but consider a situation in which the viewer is tracking motion across a scene. What happens when our eyes are moving? Think about watching that bird fly across the sky. Perhaps it is an eagle, and you want to watch it soar on the wind. As you follow its motion, your eyes and head move as the eagle flies. Indeed, while tracking motion, your eyes make movements (called smooth-pursuit move- ments) to keep the image of the eagle on the foveae of your retinae (Figure 8.9). In another setting, consider a tennis player tracking a tennis ball as it approaches his racket. His eyes track the movement of the ball from the moment her opponent hits the ball as it approaches his side of the court. His eyes, head, and body must be in continual movement to maintain alignment with the speeding tennis ball (Figure 8.10). This continual tracking of move- ment allows another method of detecting direction and speed of motion. This brings us to the corollary discharge theory, an important concept in under- standing how our visual system detects and tracks motion. The corollary discharge theory states that the feedback we get from our eye muscles as our eyes track an object FIGURE 8.8 Reichardt Detectors A simple neural network for detecting motion. The signals arrive at neuron M below receptors A and B at the same time, from a neuron with a receptive field at P1 and then from a neuron with a receptive field at P2. Depending on the time course of arrival, neuron M will determine motion and speed. A B P1 P2 (a) M Time 1 P1 P2 Neural signal M Time 2 A B P1 P2 M Time 3 Receptive Field Retinal Detectors A B P1 P2 M Time 4 Motion- Sensitive Neurons A B P1 P2 (b) M Time 1 A B A B P1 P2 Delay in signal M Time 2 A B P1 P2 M Time 3 Receptive Field Retinal Detectors A B P1 P2 M Time 4 Motion- Sensitive Neurons No neural signal FIGURE 8.9 Smooth-Pursuit Eye Movements As the bird soars across the sky, smooth-pursuit eye movements allow you to track the motion of the bird. Smooth-pursuit movements are possible only in the presence of movement. © iStockphoto.com /PeopleIm ages© iS to ck ph ot o. co m /m ic ha el m ill Corollary discharge theory: the theory that the feedback we get from our eye muscles as our eyes track an object is important to the perception of motion 235 Chapter 8: Movement and Action is important to the perception of motion. Corollary dis- charge theory starts off quite similar to the Reichardt detector model, with neurons that are sensitive to movement across the retina. However, when we wish to track an object such as a tennis ball, a command signal must be sent from the brain to the muscles that control eye movements. Corollary discharge theory states that, in addition to the muscles of the eye, this signal will be sent to areas of the brain responsible for motion detection. This signal, or corollary discharge, provides the brain with updated information about the locations and speeds of moving objects (Sommer & Wurtz, 2008). In other words, we use the fact that we are moving our eyes and heads as a means to detect motion in the world. It is as if our visual systems were interpreting the situation as follows: We are making smooth-pursuit movements with our eyes; therefore, we must be tracking motion. This logic then allows the feed- back from the eye movements to assist our visual system in seeing and timing the motion. One source of evidence for the corollary discharge theory comes from the movement of afterimages. If you look directly at a bright light and then close your eyes, you will see an afterimage of that light. When you move your eyes, the afterimage stays on the exact same part of the retina. However, you sense motion. Why is this the case, if the afterimage has not moved across your visual field? It must be because of feedback from a corollary discharge loop. Similarly, experiments have shown that people can keep their eyes fixated on a central point even when their eyes are being physically pushed. Here too, even though there is no movement of objects across the visual field, we sense motion, stemming from a corollary discharge feedback system (Bridgeman & Stark, 1991). For a demonstration of corollary discharge and how afterimages move when the eyes move, see ISLE 8.8. Physiological evidence for this theory comes from the discovery of real motion neu- rons, which respond only to movements of objects but not movements of eyes. To deter- mine the difference, these neurons must first get feedback from eye movement signals so that they can “know” the difference. Experiments with monkeys show that these real motion neurons in the extrastriate cortex respond when an object is moving across the monkeys’ visual field but not when their eyes are moving and the object is stationary (Galletti & Fattori, 2003). Corollary discharge, therefore, allows our visual system to integrate the movements of objects with our own basic motility. The corollary discharge theory means that one of the cues to detecting motion is the movement of our own eyes. Thus, it is important to understand how and why our eyes move in order to understand the nature of motion perception. We turn our attention next to eye movements. Eye Movements Our visual system can perceive objects in the world that move, but, of course, we move as well. We can move our eyes, our heads, and our bodies. We can track motion just by moving our eyes, or we can follow the direction of a moving object by walking or running. In this section, we are concerned with eye movements that direct our eyes from one location to another in the visual world. This contrasts with the discussion of eye movements with vergence, discussed in Chapter 7. Vergence refers to the eye FIGURE 8.10 Using Smooth-Pursuit Eye Movements While playing a sport such a tennis, our eyes and head move to track the movement of the ball using smooth-pursuit eye movements. © iStockphoto.com /PabloBenitezLope ISLE 8.8 Corollary Discharge 236 Sensation and Perception movements that cause our eyes to resolve a particular object or bring it into focus. Here we are concerned with changing our gaze from one object to another or tracking a moving object. We have already discussed how feedback from eye movements helps our visual system sort out different forms of motion. Eye movements serve a great many other functions as well, the most significant being the ability to direct the gaze of our foveae on whatever stimulus in the world attracts our attention. Eye movements allow us, for example, to read. We make small discrete movements along the page to ensure that our foveae are directed at whatever it is we want to read right now. When we watch moving print, such as the credits at the end of a movie, we make small eye movements from left to right and up and down. There is a simulation of eye movements on ISLE 8.9. Saccades Saccades are the most common eye movements. They are used to look from one object to another. Saccades are very rapid—we make them in less than 50 milliseconds (ms), and we can make approximately three saccades each second. In other circumstances, we may need up to 1 second to plan a saccade, but once initiated, the movement is very rapid. What characterizes saccades is the quick jump from the focus on one object to another. When you read, you make a series of saccades along the page. When you look up from what you are reading to see who is looking in the window, you make a saccade from your book to the window. Then, when you change your gaze from looking at a stationary tree to your car parked outside your house, you are also making a saccade. Interestingly, vision is suppressed during the actual movement. That is, during the 50 ms it takes to make the eye movement, we essentially cannot see anything new. Only when our saccade lands at the new location do we again see. Smooth-Pursuit Eye Movements Smooth-pursuit eye movements are the voluntary movements we use to track moving objects. We can make smooth-pursuit eye movements only when there is an actual mov- ing object in the environment. In an environment without movement, we can make only saccades. Feedback from our visual system, however, allows our eyes to continuously follow motion. Thus, our eyes move gradually as we follow an airplane across the sky or a tennis ball back and forth across a tennis court. TEST YOUR KNOWLEDGE 1. What are Reichardt detectors? What aspect of motion do they account for? 2. What are the differences between saccades and smooth-pursuit eye movements? MT: The Movement Area of the Brain MT, also known as V5, is an area in the occipital cortex critical to motion perception. MT stands for “medial temporal,” because MT is adjacent to the medial temporal lobe, although it is itself within the extrastriatal areas of the occipital lobe. In Figure 8.11, you can see where MT and other important motion perception areas are located in the brain. Recall that extrastriate refers to an area in the occipital lobe, but not in V1. MT receives input from both V1 and V2, as well as from the superior colliculus, an important area in controlling eye movements. From the research we review shortly, we know that MT is sensitive to both direction of motion and speed of motion. In contrast, MT neurons are not tuned to other perceptual characteristics, such as color or orientation. We also know that it integrates motion-sensitive cells from across V1 to ISLE 8.9 Eye Movements 237 Chapter 8: Movement and Action detect large-scale motion (such as a walking person). We can consider MT a region of the brain critical to the visual perception of motion. Some of the most important work on the func- tion of MT was done by the neurophysiologist William Newsome and his colleagues (Hedges et al., 2011; Newsome, Shadlen, Zohary, Britten, & Movshon, 1995). They were interested in how and why neurons in MT respond to motion, particularly larger scale motion, such as a dog running in a park. They set out to understand motion perception in rhesus monkeys, which have visual brains similar to our own. We now describe some of their experimental work linking MT to motion percep- tion in rhesus monkeys. The logic of the experiment is as follows. Newsome and his team trained rhesus mon- keys to respond to particular forms of motion with spe- cific behaviors. That is, if a visual pattern was moving in one direction, the monkeys made one response, but if the visual pattern was moving in the opposite direction, the monkeys made another response. Once the monkeys were trained, Newsome could lesion MT in the mon- keys’ brains and see what behavioral deficits the monkeys developed or, somewhat more humanely, conduct single-cell recordings of individual cells within MT while the monkeys were watching movement. In an important first experiment, Newsome and Paré (1988) trained monkeys to respond to a pattern of dots all moving in the same direction (Figure 8.12). However, they added an interesting manipulation. In some trials, all of the dots were moving in the same direction (100%), but in other conditions, only 50% or 20% of the dots were moving in sync, whereas the rest of the dots were moving randomly. As such, in these conditions, no dot, in and of itself, could determine the large-scale motion seen in the pattern, as any particular dot might just be going in the opposite direction. However, clearly our visual systems must be tuned to this kind of motion. (Fish swimming in a school or cattle in a herd provide real-world examples of this situation.) In essence, this experiment was looking at the physiological basis of the gestalt principle of common fate. Do we see motion when a pattern of dots is moving in a particular direction, even if an individual dot is not? The visual system must pull information across many dots to determine the pattern of motion in the 50% and 20% conditions (see Figure 8.12). By the time the monkeys were fully trained, they could detect the general pattern of motion even when only 3% of the dots were moving in the same direction (and the rest were moving randomly). That is, MT neurons integrate across any number of local detectors to infer the general pat- tern of motion in the display. You can see these patterns in motion on ISLE 8.10. Newsome and Paré (1988) then sur- gically lesioned MT in these monkeys. Following recovery from surgery, the mon- keys’ performance on the direction-of-mo- tion task was severely impaired. However, the monkeys were not impaired at visual Primary visual cortex (V1) Middle temporal and middle superior temporal areas (V5) Superior temporal sulcusLateral fissure Lingual gyrus (V3) Motion-processing parietal cortex Postcentral sulcus Central sulcus FIGURE 8.11 Motion Areas in the Brain In this map of the brain, you can see V1 at the back of the occipital lobe. MT (also known as V5) is anterior in the occipital lobe, located just above the temporal lobe. This map also shows motion-sensitive areas in the parietal lobe. ISLE 8.10 Correlated Motion 100% 50% 20% FIGURE 8.12 Stimuli Used in Newsome and Paré’s (1988) Experiment In the 100% condition, all the dots were moving in the same direction. In the 50% condition, half the dots were moving in a consistent direction. And in the 20% condition, only one fifth of the dots were moving in a consistent direction. 238 Sensation and Perception tasks that involved stationary objects—only motion tasks were affected. Thus, for exam- ple, color identification was fine, but the direction-of-motion task showed strong decre- ments in performance. Thus, because these monkeys had been able to do motion tasks prior to the lesioning but were impaired afterward, Newsome and Paré concluded that MT is involved in this kind of motion discrimination. In another experiment (Newsome, Britten, & Movshon, 1989), Newsome and his colleagues found through single-cell recording that MT neurons fired more strongly the more coherent or correlated the motion was. That is, there was more activity in MT when all the dots were moving in the same direction than when only 20% of the dots were moving in the same direction. This was the case even though the 20% motion is salient when humans view it and presumably similarly salient to the monkeys. Thus, these studies provided strong evidence that MT is not just a motion area but one that is sensitive to larger scale motion. There have also been studies using neuroimaging that confirmed the role of MT in humans. For example, fMRI studies with human participants have confirmed the role of MT in such large-scale movement, similar to Newsome’s studies. In humans, as with monkeys, MT responds more to moving stimuli than to stationary stimuli, and more to complex movement than to simple movement. For example, Weigelt, Singer, and Kohler (2013) looked at the human MT by recording activity with fMRI technology. Weigelt et al. showed participants moving stimuli. In some tri- als, participants attended to the locations of certain objects, whereas in other trials, participants attended to the direction of motion of certain objects. In this case, the stimuli were identical—the difference was what the person was attending to, either the locations of objects or the direction of motion. Weigelt et al. found that attention to motion resulted in a bigger response in MT than did attention to location. In con- trast, they found no difference in the neural responses in V1 between the two attention conditions. Thus, as in monkeys, the human MT is responsive to large-scale motion. Weigelt et al.’s results can be seen in Figure 8.13. Weigelt and her group have also found evidence that MT is active during visual imagery that involves motion (Kaas, Weigelt, Roebroeck, Kohler, & Muckli, 2010). That is, even when there is no actual movement present, but participants are imagining moving objects, MT is active. In the study, participants were asked to imagine objects moving without actually seeing the motion of those objects. They compared this with a visual imagery condition that did not involve motion as well as an auditory imagery task. Relative to the control conditions, when participants were engaged in motion visual imagery, larger responses were recorded from MT. These results can be seen in Figure 8.14. Here too, we can see how important MT is to higher order aspects of motion perception. Last, we turn to neuropsychology and its role in understanding the function of area MT. Neuropsychology provides us with some of the most convincing and, at the same time, bizarre evidence for the role of MT in motion perception. When MT is damaged, FIGURE 8.13 Weigelt et al.’s (2013) Results In this fMRI study, Weigelt et al. found that activity in MT was associated with attention to movement. In these fMRI pictures, the activity in MT is highlighted in color. The results are averaged across 15 participants. 239 Chapter 8: Movement and Action a condition called akinetopsia or motion blindness can arise. Akinetopsia is a rare condition in which a patient is unable to detect motion despite intact visual perception of stationary stimuli. Patients with this disorder no longer see objects in motion. Rather, their perception resembles a series of still photographs moving one to the next. It occurs without any loss of object recognition, color vision, or visual acuity, but nonetheless, it can have profound effects on a person’s vision and ability to function in a visual world. What would the world be like for someone with no perception of visual motion? A potential metaphor in normal experience is being in a dark room illu- minated by a stroboscope flickering slowly, as might occur in some dance clubs. The room is dark except for brief flashes in which it is illuminated. Under these conditions, normal people may feel as if they have motion blindness. Several patients have been tested and shown to have motion blindness. One of the most widely tested is a patient labeled L.M. (Zihl, von Cramon, & Mai, 1983). Patient L.M. described the difficulty of pouring water into a glass, because she could not see the movement of the water from the pitcher into the glass and therefore could not tell exactly when the glass was full. Similarly, crossing a street is difficult for L.M., because she does not know if her current image of the world is recent and therefore how close the cars actually are to passing in front of her. From this testimony of L.M., we can see how important it is to be able to perceive motion as it is happening. Schenk, Mai, Ditterich, and Zihl (2000) showed that a patient with akinetopsia had difficulties with simple action tasks as well. When instructed to grasp moving objects, the motion-blind patient took more time and was less accurate at grasping objects than age-matched controls. Thus, the deficit in motion blindness extends to control of the motor system as well as the profound perception deficits it causes. In another rare but documented case, epileptic seizures induced temporary akinetopsia in at least one patient (Sakurai, Kurita, Takeda, Shiraishi, & Kusumi, 2013). In summary, damage to MT in human patients can cause a terrible motion perception deficit. Interestingly, akinetopsia can be simulated in normal human brains. A technique called TMS allows researchers to temporarily disable selective areas of the human brain. TMS stands for “transcranial magnetic stimulation.” In some cases, it can increase activity in the area of the brain to which it is applied, but it also can decrease activity. It depends on the kind of stimulation that is applied and what behavior is being measured. Interestingly, when TMS is applied to area MT, temporary akinetopsia results (Beckers & Zeki, 1995). These participants will experience a temporary inability to perceive motion, which, fortu- nately, wears off shortly after the TMS is no longer being applied. Motion Aftereffects One of the illusions we see quite commonly in everyday life is the motion aftereffect. Motion aftereffects are also called the “waterfall illusion” because they occur when we stare at a waterfall (Figure 8.15). After we have watched the falling of the water for about a minute, if we look at a blank surface, such as a white wall, we will get a sense of motion going upward, a. Individual hMT/V5+ regions of interest b. Regions in�uencing left hMT/V5+ during motion imagery 1. Left anterior caudate nucleus 2. Left inferior parietal lobule ImageryImagery y = −38y = −10 t(11) = 2.2 p < 0.05 ControlControl FIGURE 8.14 fMRI Results From Kaas et al. (2010) These diagrams illustrate how MT is activated when people imagine perceived motion, particularly 8.14a. Akinetopsia (motion blindness): a rare condition in which an individual is unable to detect motion despite intact visual perception of stationary stimuli, caused by damage to area MT 240 Sensation and Perception that is, in the opposite direction of the falling water. Perhaps today, the most common experience of motion aftereffects comes when we watch the credits at the end of a movie. If you want to know where the movie was filmed, for exam- ple, you may dutifully watch the names of the actors and actresses scroll up the screen (apparent motion, not real motion), the stunt staff, and the camera staff, until finally, the credits reach the spot where they tell what town and in what state or country the movie was filmed. At this point, you may pause your DVD player or your mobile device’s video player to read the information about where the movie was filmed. When you pause the apparent motion, you may get a definite sense of the words on the screen moving upward after watching them scroll downward for a couple of minutes. This illusion of a nonmoving surface moving opposite to the movement just watched is called the motion aftereffect or waterfall illusion. For a demonstration of motion aftereffects, go to ISLE 8.11. We can define a motion aftereffect as a motion-based visual illusion in which a sta- tionary object is seen as moving in the opposite direction of real or apparent motion just observed. Aristotle wrote about the waterfall illusion around 350 BCE. He described watching the water tumbling down the falls for several entranced minutes. He then fixed his gaze on a stationary object and experienced the illusion of that object’s moving upward, opposite the direction of the waterfall (see Figure 8.15). The motion aftereffect suggests that motion neurons in the occipital cortex may have an opponent system, similar to that for color vision. Because we see movement opposite the direction of the motion we were just watching, the motion aftereffect suggests that movement in one direction is linked to movement in the opposite direc- tion. Thus, upward and downward motion is an example of motion contrast, just as red and green are color contrasts. Consider two motion-sensitive neurons in area MT of the occipital cortex. One neuron, N-l, is sensitive to motion in the upward direction, whereas N-r is sensitive to downward motion. Activity in N-l may prompt inhibition in N-r, but when N-l is no longer activated and the inhibition is turned off, we will sense the downward motion that has been inhibited. Of course, neurons are sensitive to many features in addition to the one under discussion. Thus, N-l might also have particular brightness, orientation, and speed preferences. For example, N-l might be stimulated by a dim object moving in the neuron’s preferred direction or by a very bright object moving in a less than optimal direction for that neuron. On the basis of this cell’s response alone, we cannot tell the difference between these two possibilities. However, if N-r is tuned to the opposite direction of motion, but perhaps other features, the inhibition it gets as a function of movement in the opposite direc- tion is enough to disambiguate the N-l stimulation. Thus, motion aftereffects are an illusory consequence of a system that allows neurons to pick out specific directions of motion. TEST YOUR KNOWLEDGE 1. What were Newsome and Paré (1988) interested in? What methodology did they use to address this issue? 2. What is akinetopsia, and what does it result from? FIGURE 8.15 The Waterfall Illusion: Motion Aftereffects A beautiful waterfall in Tennessee. If we stare at the downward motion of a real waterfall, we can see the aftereffect when we look at another surface afterward. © iS to ck ph ot o. co m /B am bi G ISLE 8.11 Motion Aftereffect Motion aftereffect: a motion- based visual illusion in which a stationary object is seen as moving in the opposite direction of real or apparent motion just observed 241 Chapter 8: Movement and Action FORM PERCEPTION AND BIOLOGICAL MOTION 8.3 Discuss how we infer form from the perception of biological motion. Scared animals instinctively know to stay perfectly still, especially those species that wear camouflage on their skin or fur. Think of an arctic hare hiding in the snow from wolves or polar bears (Figure 8.16). Its white fur allows it to blend in perfectly with the winter snow. However, keeping still is also critical to an animal or a person attempting to hide via camouflage. Once prey is in motion, predators can detect the motion, despite the camouflage. Indeed, we demonstrate in this section that we are often able to detect form directly from the perception of motion when motion cues are the only cues to what an object is. See ISLE 8.12. Motion perception provides us with information about what object we are observing. Think about someone you know with a distinctive style of walking. You may know some- one who always appears to have a hop in her step, and you may know someone else who seems to be trudging around all the time. These motion cues may help you recognize that person. Consider running into one of these people at night, when you cannot get a good view of the person’s face. Detecting the style of walking may allow you to recognize the per- son. An ingenious technique allows researchers to assess whether these assertions are true. Johansson (1973) developed a method called the point-light walker display to examine this phenomenon. A point-light walker display is one in which small lights are attached to the body of a person or other animal, which is then filmed moving in an otherwise completely dark environment. In Johansson’s experiment, small lights were placed on a person’s body, including the wrists, elbows, shoulders, ankles, knees, and hips. He then video recorded the person walking, running, and dancing in total darkness. Thus, the only thing visible to viewers were the lights, and not the person’s body (Figure 8.17). It is absolutely necessary to view the demonstration and experiment on this topic on ISLE 8.13a. You can also go to http://www.biomotionlab.ca/Demos/ BMLwalker.html (also linked to in ISLE 8.13a) to view a demonstration. When Johansson (1973) showed participants still photographs of the light patterns, the participants were unable to extract any form information from the photographs. However, when participants viewed videos of the light patterns, they were able to detect the human form when it was walking or running. Moreover, the participants were able to distinguish between forms of motion, including walking and running. Other studies using similar methods have shown that human observers can distinguish between dif- ferent animals on the basis of the patterns of their movements in point-light displays (Blake & Shiffrar, 2007). In addition, human participants can judge age, gender, and a person’s emotional state solely on the basis of point-light displays (Blake & Shiffrar, 2007). Finally, human observers can learn distinctive walking patterns of individuals in point-light displays and later recognize those individuals when they see them moving in a normal video (Westhoff & Troje, 2007). Point-light displays also allow human beings across cultures to infer the emotional state of the person in the display (Parkinson, Walker, Memmi, & Wheatley, 2017). Thus, these studies support the notion that pat- terns of motion can allow us to see or infer the presence of form. It is also possible to disrupt our usual good perception in point-light displays. Troje and Westhoff (2006), for example, showed videos of biological motion through point-light displays of people walking, but the videos were inverted. Under these FIGURE 8.16 A Well- Camouflaged Arctic Hare As long as it stays still, it is hard to detect. © iStockphoto.com/photos_martYmage ISLE 8.12 Structure From Motion ISLE 8.13 Biological Motion Point-light walker display: an experiment in which small lights are attached to the body of a person or an animal, which is then filmed moving in an otherwise completely dark environment 242 Sensation and Perception circumstances, it was very difficult for participants to detect the biological motion. Breaking down the coordination between the lights on various parts of the body can also interfere with perception under these circumstances (Troje, 2013). Interestingly, children with autism are impaired at inferring form in point-light displays relative to matched con- trols (Swettenham et al., 2013). Thus, variables or situations that affect our perception of patterns can render form-from- motion extraction more difficult. To see some of these exceptions, go to ISLE 8.13b. As with most successful experimental paradigms in sen- sation and perception (and all of psychology), the point-light display procedure has now been used while participants are monitored by fMRI technology. The perception of biological motion in these point-light displays is associated with activ- ity in the posterior superior temporal sulcus in the temporal lobe (Troje, 2013). In studies using point-light displays in which neural activities have been monitored by fMRI, there is more activity in the posterior superior temporal sulcus for biological motion than nonbiological motion. Thus, this region is not motion specific—it does not rise above base- line activity for nonbiological motion. Rather, it seems to be involved in the recognition of particular forms of motion, consistent with the idea that we can use motion perception to infer forms. In another study, Grossman, Battelli, and Pascual-Leone (2005) interfered with the posterior supe- rior temporal sulcus with a pulse from a TMS device. While under the influence of TMS, normal individuals showed a severe deficit in the perception of biological motion. As you can see, this methodology has been enormously successful in examining higher order levels of movement. TEST YOUR KNOWLEDGE 1. What information does motion perception provide that is not present in static images? 2. What is a point-light walker display? What does it tell us about how motion contributes to form perception? ACTION 8.4 Summarize the concept of an affordance. Young cats existed for millions of years before the invention of the laser pointer. In those bygone days, kittens found lots of things to play with. But with the laser pointer, the young cat has found the ultimate human–cat interface. Even though laser pointers were invented for less exciting reasons, including allowing professors to point to over- head screens, laser pointers afford playing in young cats. And for lazy cat owners, laser pointers afford play sessions with their cats without their having to get off the couch (Figure 8.18). We will see that the concept of an affordance is critical in understanding Person wearing point lights Point-light walker display FIGURE 8.17 Point-Light Walker Display Lights are attached to a person’s joints. The person is filmed walking or running in a completely dark room except for the lights on his joints. Then the point-light displays are shown to participants. In general, most participants can determine the type of and direction of motion from these point-light displays alone. We can even determine the gender of the walker from the lights alone. 243 Chapter 8: Movement and Action the theory of the interplay between perception and action that we are about to develop. If you are a bit tired of studying and need a laugh, you can watch Dr. Krantz playing the laser game with his cats on ISLE 8.14. An affordance is the information in the visual world that specifies how that information can be used. For example, seeing a piano affords playing music. Seeing an elevator affords being able to go up or down in a tall building. For a track athlete, seeing a hurdle affords jump- ing over it. And for a kitten, seeing the laser light of a laser pointer affords high-speed chases across the floor. According to the prominent vision theorist J. J. Gibson (1979), perception is about finding these affordances in the world. Perception, according to Gibson, is not about whether a person can detect a very dim spot in an oth- erwise completely dark room, but rather about percep- tion in everyday life and how we use perception to guide action. Animals, including people, see in an ever-changing world while constantly moving themselves. As such, Gibson argued, we should pay more attention to the complexity of stimuli in vision. Moreover, he argued, the goal of perception itself is to afford action, that is, movement of the organism. Affordances mean that perception is determined partially by meaning or function. We see objects as stimuli with particular characteristics not just of color, depth, width, size, and shape but also of function. An object that has no function may also not draw our attention, whereas we seek out objects that we need at the moment. Not seeing where your eyeglasses are, for example, may elicit some search behavior, but you are not likely to search for an object with no apparent use. Similarly, we may judge bean- bags to be more similar to regular chairs than we might expect on the basis of their similarity in shape. This is because beanbags and chairs share a common function: They afford sitting in comfort. Gibson (1979) emphasized that function drives perception. One of Gibson’s chief contributions was to the understanding of optic flow. We discussed optic flow in Chapter 7 as one of the motion cues to depth. We defined optic flow as a motion depth cue that refers to the relative motions of objects as the observer moves forward or backward in a scene. In this chapter, we focus our attention on how we perceive motion in optic flow patterns and how they influence our action in motion. To see an optic flow pattern of driving along a road or landing a plane, go to ISLE 8.15 (also see Figure 8.19). FIGURE 8.18 For a young cat, a laser pointer affords hours of fun. © iStockphoto.com /borzyw oj ISLE 8.14 Cat and Laser Pointer Affordance: information in the visual world that specifies how that information can be used FIGURE 8.19 Optic Flow As we move in one direction, objects appear to get bigger as they approach us. In addition to this being a cue for depth, it also indicates motion. A driver can judge her speed by the rate at which objects are flowing toward her. © C an S to ck P ho to In c. /n ac ro ba ISLE 8.15 Optic Flow 244 Sensation and Perception Optic flow provides information about distance and consequently can be used to aid in our own movement. As we discussed in the previous chapter, objects closer to us seem to move faster past us, whereas more distant objects appear to move slowly as we move toward them. This pattern is called the gradient of flow. In simulations, if we increase the speed at which the observer is moving, this gradient of flow accelerates. If the speed of the observer decreases, the gradient of flow lessens. The horizon or destination point is the distant point at which the gradient of flow comes to zero. This point is called the focus of expansion. As we voluntarily change directions in an optic flow pattern, this focus of expansion will shift. In this section, we are interested in how action and perception interact. At a sim- ple level, knowing where you are going means that you can assess if you are actually going in the intended direction. If your goal is to pilot your small Cessna plane from Fort Lauderdale, Florida, to Bimini, in the Bahamas, decisions about slight deviations in course will occur with respect to both the goal and your present position. A strong wind starts blowing out of the east, slowing your speed, and you may seek to fly at a different altitude. In a virtual reality experiment, Warren, Kay, Zosh, Duchon, and Sahuc (2001) controlled the optic flow pattern that participants were viewing while walking in a specially designed room. Participants were asked to walk toward a goal: a red line visible in the virtual reality setup. When Warren et al. slowed or sped up the optic flow pattern, participants adjusted their walking to be consistent with their goal of making it to the red line. Thus, optic flow patterns can change the actions we are engaged in. This result is consistent with many others demonstrating that optic flow plays an important role in adjusting our movements as well as allowing us to perceive depth (Chrastil, 2013). In the modern world, driving a car is an important and frequent activity. Ensuring safe driving should be—even if it is not—a national concern. Thus, examining how drivers use optic flow is also relevant. Do drivers use optic flow patterns to ensure that they are driving safely and in the correct direction? Some research suggests that optic flow is one way in which drivers maintain their positions on the road. For example, Kandil, Rotter, and Lappe (2010) were interested in what information drivers use when driving on bends or turns. While rounding a turn, which is quite normal for drivers, the focus of expansion will not be obvious to the driver’s visual system. Thus, Kandil and colleagues were interested in what other cues drivers use to keep their cars on the road and going in the correct direction while driving on curves. In their exper- iment, people drove real cars on real roads, while their eye movements were being recorded. The investigators recorded these eye movements as drivers drove through a series of bends. They found that drivers kept their eyes on the focus of expansion during straight-ahead driving. But as drivers approached bends, they took their eyes off the focus of expansion and instead directed their eyes to the straight road segments that were coming up. This focus on the goal of rounding the turn—the straight stretch of road ahead—allowed the drivers to direct their cars’ movement in a smooth pattern around the bend. New drivers tend to focus on the curve itself rather than the straight stretch ahead, leading to more adjustments and less smooth driving. Thus, optic flow information is one source, but not the only source, of information drivers use to adjust their motion. Visually Guided Eye Movements The anatomical bases for the interaction between action and perception emerge from the parietal lobe. The parietal lobe appears to have critical networks for integrating Gradient of flow: the difference in the perception of the speeds of objects moving past us in an optic flow display Focus of expansion: the destination point in an optic flow display, from which point perceived motion derives 245 Chapter 8: Movement and Action muscle systems, somatosensory systems, and the visual sys- tem. This interaction allows the smooth transition from visual perception to guided action. The parietal lobe itself is divided by the postcentral sulcus, which divides the anterior parietal lobe from the posterior parietal lobe (Figure 8.20). The anterior parietal lobe is concerned with the somato- sensory system, a topic for a later chapter. However, the posterior parietal lobe contains regions that are specific to visually guided action (Snyder, Batista, & Anderson, 2000). Studies with monkeys reveal an area known as the lateral intraparietal (LIP) area, which is involved in the control of eye movements. There appears to be an equivalent area in human brains. (You can see where this region is located in Figure 8.20.) Colby, Duhamel, and Goldberg (1996) performed some interesting experiments on the role of the LIP area using sin- gle-cell recording with rhesus monkeys. In these experiments, a monkey is required to hold fixation on a central point (Figure 8.21). Once the monkey has started to maintain fix- ation, the experimenters flash another point at the near periphery relative to the fixation point. The monkeys are trained to not move their eyes to this new point until after it disappears. When the point disappears, the monkeys make a saccade to this sec- ond point, even though the light is no longer pres- ent. The researchers were interested in what the LIP area was doing during the original appearance of the signal at the periphery and when the monkey actu- ally made the eye movement. Colby et al. found that neurons within the LIP area, which had a receptive field corresponding to the area where the second spot appeared, became active as soon as the spot appeared on the screen. Activity in the LIP area then continued as the monkey made its saccade to this spot and then diminished afterward. Thus, the LIP area of the pari- etal lobe is involved with both the anticipation of the eye movement and the eye movement itself. Studies using fMRI with humans have found an equivalent area in the human brain that also shows activity that reflects intended eye movements (Quiroga, Snyder, Batista, Cui, & Andersen, 2006). Visually Guided Grasping As we have emphasized in this chapter, the visual system functions to provide us with information, which we can then use to guide our actions. Toward this end, our visual systems and our motor systems have coevolved to allow us precise control over move- ment on the basis of continual feedback from our visual system. This aspect of vision– action interaction can be seen in research on the control of grasping in humans and other primates. Primate hands, especially human primate hands, are instruments of fine Postcentral sulcusCentral sulcus Superior parietal lobe Intraparietal sulcus Occipital lobe Inferior parietal lobe Frontal lobe Lateral fissure Temporal lobe A nt er io r pa rie ta l l ob e Po ste rior parietal lobe M IP LIP AIP FIGURE 8.20 The Parietal Lobe Serves Important Functions in the Perception of Action Studies with monkeys have revealed an area known as the lateral intraparietal (LIP) area, which is involved in the control of eye movements. This area is marked on the human brain shown here. Lateral intraparietal (LIP) area: an area of the primate parietal cortex involved in the control of eye movements Receptive field of LIP neuron Target spot Monkey looks at fixation point Target spot appears Time = t 0 Time = t1 Target spot disappears Time = t 2 Saccade Fixation point disappears Time = t 3 Fixation point Direction of gaze FIGURE 8.21 Experiment Showing the Function of the LIP Region of the Parietal Lobe Monkeys are expected to keep their eyes on a central fixation point. The researchers can then look at the receptive field of a particular neuron in the LIP area. A target then appears in the receptive field of that neuron. When the target disappears, the monkey must make a saccade to that location. Activity in the LIP area precedes the saccade. 246 Sensation and Perception motor control. Think of all the tasks you do that require fine motor control of your hands, from typing on your computer keyboard to knitting a sweater to changing the spark plugs on your car. Each of these tasks requires fine motor control of the hand combined with complex feedback from the visual system. If you refer back to Figure 8.20, you will see an area of the posterior parietal lobe identified as the medial intraparietal (MIP) area. This area of the parietal lobe is involved in the planning and control of reaching movements of the arms. The MIP area seems to be critical in guiding the movement of the arms toward a visu- ally selected location. MIP area neurons also seem to be involved in goal-directed arm movements (Swaminathan, Masse, & Freedman, 2013). Nearby is the anterior intraparietal (AIP) area, a region of the posterior parietal lobe involved in the act of grasping. Once your arms are close enough to an object that your fingers can start picking the item up or grasping it, activation shifts from MIP area neurons to AIP area neurons. Gallivan, McLean, Flanagan, and Culham (2013), using fMRI, showed that the AIP area was active during visually guided movements to grasp at specific objects. Moreover, like the LIP area, which controls eye movements, the AIP area was also active during the planning of grasping motions. Gallivan et al. had participants plan to grasp either small or large objects, which would require different hand movements. Activity occurred in the AIP area before the participants were given instructions to actually move their hands. These experiments elegantly demonstrate the interaction between visual and motor systems in subserving percep- tually guided action. TEST YOUR KNOWLEDGE 1. What is meant by the term optic flow? What does it tell us about how we use visual information to inform our own motion? 2. What is the lateral intraparietal area? Where is it in the brain, and what function is it linked to? What is the function of the anterior intraparietal (AIP) area and the medial intraparietal (MIP) area? Medial intraparietal (MIP) area: an area of the posterior parietal lobe involved in the planning and control of reaching movements of the arms Anterior intraparietal (AIP) area: a region of the posterior parietal lobe involved in the act of grasping EXPLORATION: Motion Illusions Illusions can be thought of as glitches in the system. At some level, they are errors—we misperceive what is actu- ally out there. In a perfect visual system, there would be no illusions. But our visual system, however effective it is, is not perfect—its mechanisms occasionally lead to errors, which we call illusions. We have made the point that illu- sions often occur because of unusual situations that our visual systems were not designed to handle, but that our visual systems are very good at handling natural percep- tion. While viewing an illusion, we do not see what is really present, and we do see what is not. However, illusions can be fun. We often pay to see professional magicians enter- tain us with illusions. And more important, illusions can reveal to us how our perceptual systems work. By finding when and where these mechanisms fail, we can figure out how the perceptual system is working in the first place. Inspection of the various motion-based illusions is no exception to this rule. Take a look at Figure 8.22. There is no actual movement in this figure—how could there be, especially if you are looking at a piece of paper in a textbook? Yet the illu- sion of movement is powerful. When you look at the center of one of the snakes, that snake appears station- ary, but all the others in the figure appear to be rotating. Dr. Kitaoka’s website (http://www.ritsumei.ac.jp/~akitaoka/ index-e.html) shows many variations on this theme. We have already discussed the famous waterfall illusion or 247 Chapter 8: Movement and Action motion aftereffect. This adaptation effect tells us much about how neurons are organized in area MT of the occip- ital cortex. It also tells us about the nature of saturation of motion channels. Watching movement at a particular speed exhausts the neurons tuned to motion at that speed, leading to the characteristic sense of motion going the opposite direction when we look away from a steadily moving object. In this Exploration section, we address several interesting and important motion illusions and discuss what they tell us about motion perception. Keep your access to the ISLE system handy. As with all illusions, seeing them is usually necessary to understand them, so reading this section without viewing the illusions is not advised. And because a static book cannot show move- ment, you will certainly need to access your electronic resources for this section. Illusion 1: Rotating Snakes The illusion of rotating snakes (see Figure 8.22) was developed by Akiyoshi Kitaoka (2005), a psychologist in Kyoto, Japan. Because this illusion creates an illusion of motion from a static display, you can see the motion in Figure 8.22. It is also in ISLE 8.16. You can also visit Dr. Kitaoka’s website (http://www.ritsumei.ac.jp/~akitaoka/ index-e.html) to view not only this illusion but many other striking motion illusions that occur with static dis- plays. The rotating snakes illusion is based partially on the earlier Fraser-Wilcox spiral staircase illusion (Fraser & Wilcox, 1979), which you can see on ISLE 8.17. Kitaoka, however, has multiplied the effect. If you look at the center of any particular “snake,” you can see that the snake you are looking at is stationary. However, you will notice that the rest of the “snakes” are in motion. The motion is illusory—the figure is completely static. You can test this by changing your gaze and looking at another “snake.” Now the new snake is stationary and the one you were just looking at appears to move. And if you are looking at the illusion on paper rather than on a computer screen, you know it is impossible for that paper to move in the way you are perceiving motion. If you are viewing the illusion on a computer screen, con- sider printing out the illusion to convince yourself that the motion is illusory. The explanation of the motion induction revolves around eye movements. As we discussed earlier, saccades are eye movements we make in which we direct our gaze from one location to a new location. Some of these saccades may involve very small changes in focus. These micro- saccades are sometimes so small that we do not even notice them as we make them. A micro-saccade can be as small as changing one’s focus from the first “hump” in the letter m to the second one. According to Otero-Millan, Macknik, and Martinez-Conde (2012), the explanation for the rotating snakes involves these micro-saccades. In their study, participants were shown the rotating snake illusion and asked to indicate when the perceived motion was at a maximum and to indicate when the figure appeared to be not rotating. The perception of motion was highly correlated with the micro-saccades, which they measured via eye tracking. Apparently, we make small saccades along the figure but do not notice them. Instead, the change of position of our eyes relative to the figure induces an illu- sion of motion. Knowing this, you can focus your gaze on the center of one snake and focus really hard on keeping your eyes still. What you will notice when you do this is that the motion tends to stop. Then move your eyes to a new location, and the snakes will start to spin again. Prior to Otero-Millan et al.’s (2012) work, it was assumed that the illusion was caused by random drift movements of the eyes. But Otero- Millan et al. showed that it was saccade eye movements that induce the perceived motion. Micro-saccades have an important function. They pre- vent visual fading during fixation. Visual fading refers to the loss of visual experience that occurs when nei- ther the visual world nor our eyes are moving. What this means is that if an image were perfectly aligned with our retina for even a short time, there would be FIGURE 8.22 The Rotating Snakes Illusion © A .Kitaoka 2003 ISLE 8.16 Rotating Snakes ISLE 8.17 Spiral Staircase 248 Sensation and Perception rapid adaptation, and the image would quickly fade. Although this never happens in the real world, in the lab, it is possible to create a stabilized image. In a stabi- lized image, whenever and however we move our eyes, the visual world moves with them. That is, no matter how we move our eyes, we are still looking at the exact same static picture. Under these conditions, unless the picture we are viewing changes, the image quickly fades (Troncoso, Macknik, & Martinez-Conde, 2008). Micro- saccades prevent this from happening in the real world but also lead to the induction of these motion illusions from static stimuli. Illusion 2: Illusory Rotation As with most illusions, you must see this one first to appre- ciate the effect and then to understand its cause. Because it involves apparent motion, you must go to either ISLE 8.18 or the website of this illusion’s creator, Dr. Stuart Anstis (http://anstislab.ucsd .edu/illusions). A static view of this illusion is shown in Figure 8.23. Anstis calls this one the “spoked wheel” illu- sion (Anstis & Rogers, 2011). When you watch the pattern in the illusion, you see what looks like a bicycle wheel. The spokes of the disk are thin gray lines. These lines do not move, nor do they change their brightness. The apparent movement comes from the wedges between the spokes. The movement comes from the clockwise rotation of the gray of the wedges. Once the illusion is set in motion, people see both the clockwise motion of the wedges and what looks like counterclockwise motion of the spokes. This motion is both apparent and illusory, because these lines do not move. That is, the spokes remain absolutely immobile, whereas the gray wedges shift position from one frame to the next. Moreover, if you stare at the pattern long enough and then look at a clear white screen or sheet of paper, you will see a subtle counterclockwise motion aftereffect. Evidently, the apparent motion of the wedges interacts with the edge detection system to create the backward motion of the spokes (Anstis & Rogers, 2011). The spoked wheel display likely arises from the inter- action of edge contrasts that the visual system is tuned to perceive and the detection of motion in the appar- ent motion display. Because the motion system is very sensitive to time, when we introduce factors that influ- ence the visual system’s ability to discriminate stimuli quickly, we can see illusions. In this case, it is likely that the different grays are processed at different speeds, with lighter grays being processed more quickly. This creates differences in the speed of processing, which, in this case, we perceive as the counterclockwise movement of the spokes. Illusion 3: The Furrow Illusion This is another illusion developed and studied by Anstis (2012). You can view this illusion either on ISLE 8.19 or on Dr. Anstis’s website. A static image of it appears in Figure 8.24. Take a minute to view the illusion before continuing to read this section. When you are looking at the static image (Figure 8.24), you see a circle with an odd pattern of gray and white inside it. You also see a small yellow circle along the top and another yellow circle in the bottom half of the larger circle. When the image is set in motion, the yellow circles move from left to right or right to left across the larger circle. When you look at the top yellow circle, it moves straight across the field from left to right and back again. However, when you look at the top yellow circle, the lower yellow circle appears to zigzag back and forth as it makes its way across the screen. But if you shift your gaze and now look at the lower circle, that circle appears to move in a straight line, and it is the top FIGURE 8.23 The Spoked Wheel Illusion When you see the static image in the figure, you see what looks like a bicycle wheel. Once the illusion is set in motion, people see the clockwise motion of the wedges, but can also see what looks like counterclockwise motion of the spokes. A ns tis w eb si te ISLE 8.18 Illusory Rotation ISLE 8.19 The Furrow Illusion 249 Chapter 8: Movement and Action yellow circle that appears to zigzag back and forth. Thus, the same image causes a different perception of motion depending on whether we are viewing it with our foveal vision or with our peripheral vision. Anstis argues that the intersections between the grating within the circle and the moving colored circles cause the illusion, revealing differ- ences in the way in which motion is perceived in the fovea and at the periphery. What do these illusions tell us about motion percep- tion? In general, we could make the case that motion perception interacts with other aspects of visual process- ing. In these illusions, we have seen how motion can be seen in nonmoving displays (rotating snakes) and that motion can be induced by the interaction of apparent movement and edge detection (illusory rotation). We also saw how motion perception can be induced when what is actually moving is our eyes (rotating snakes). Because motion perception may be different in the fovea and at the periphery, it is clear that motion perception interacts with other aspects of visual perception. We also hope you found looking at these illusions to be a “wow” experience. FIGURE 8.24 The Furrow Illusion When the image is set in motion, the yellow circles move from left to right or right to left across the larger circle. When you look at the top yellow circle, it moves straight across the field from left to right and back again. However, when you look at the top yellow circle, the lower yellow circle appears to zigzag back and forth as it makes its way across the screen. http://w w w .journalofvision.org/content/12/12/12 - via CCC APPLICATION: Motion Perception in Airplane Pilots Airplanes are a ubiquitous part of modern human life. Estimates suggest that on any given day, there are 1.7 mil- lion people on airplanes in the United States and as many as 8 million worldwide (http://www.transtats.bts.gov). Air travel is remarkably safe with an average fatality rate in the United States of less than one person per year. In the United States alone, nearly 600,000 people have licenses to pilot planes of one kind or another. In many cases, pilots can use “autopilot” and have various computer assistance, but in the end, most planes are flown by human percep- tual processes. Therefore, it is of interest to examine the processes involved in perception during flying and if there are technological adjustments that can be made to make pilots even better at avoiding danger and getting their pas- sengers safely to their destinations. Pilots and the people who train pilots must understand perception. Pilots are trained to think about the limitations imposed by our perceptual systems and what they can and cannot see, hear, or feel. They are also introduced to the malleable nature of perception, namely that pilots, like others, are subject to illusions. In some situations, if a pilot decides to continue on a current path based on erroneous information from misperceiving certain factors, such as weather conditions, runway distances, and aircraft perfor- mance, there may be situations in which accidents occur. Thus, finding out what perceptual errors pilots make and how best to correct them is paramount. Think of some of the decisions a pilot must make based on the visual perception of moving objects. First and fore- most, at take-off and landing, a pilot must judge his speed relative to the lift of the airplane. Landing at a relatively short runway means judging the angle between the air- plane and the runway to make a smooth and safe touch- down. For example, a pilot approaching a long runway 250 Sensation and Perception may perceive that her plane is too high on the approach and therefore descends too close to the ground. Pilots usu- ally have redundant systems to determine such factors as altitude, forward speed, vertical speed, direction, and so forth. These are crucial to any pilot. But pilots must also develop an ability to “feel their plane.” One such factor is judging approximate speed by watching the plane’s optic flow pattern on the ground (Yasuaki & Seno, 2015). Yasuaki and Seno have shown that topographical features on the ground can affect a pilot’s judgment of speed in the air. For example, if there are no hills or other features, a pilot may judge the ground to be moving faster than when there are hills. While flying, pilots must avoid other flying objects. Being able to determine the difference between a small object flying at low speeds nearby and a larger object flying at higher speeds farther away is an important issue and one that involves both the perception of movement and the size constancies discussed earlier. Some accidents in air travel occur when a pilot misjudges the distance between his own plane and that of another. Modern pilots, of course, have electronic traffic collision avoidance sys- tems, but pilots must also be able to make these decisions. Indeed, commercial airplanes have a system, the traffic collision avoidance system (TCAS), that monitors other aircraft near the plane and displays this information to the pilots. This display both plots the positions and rela- tive motions of other airplanes and changes the colors of other airplanes on the display if their position and motion indicate that the two planes will get too close (Livadas, Lygeros, & Lynch, 2000). Consider this situation described by Hershenson and Samuels (1999). Samuels, a professor at the U.S. Air Force Academy, was watching a plane from his office window, a not uncommon experience when working at the Air Force Academy. The plane, in particular, was doing very fast maneuvers, which gave Dr. Samuels pause. No human pilot could withstand such tight turns. This reoriented Dr. Samuels, and he realized that he was watching a smaller drone that was much closer to him rather than a larger plane at some distance. Motion-in-depth is judged by its approximate distance. An object seen moving will be perceived as slower and smaller up close but faster and larger when farther away. Thus, according to Hershenson and Samuels, when the smaller craft or drone is perceived as a larger airplane flying at a farther distance, the percep- tion will be of fast speed for that airplane. So, consider the situation of the observer. This illusion of speed occurs because the larger airplane appears to travel a greater distance in the same time that it takes the smaller craft to travel a much shorter distance. Thus, a relatively small plane subjectively appears to be a much larger airplane. In addition to appearing like a larger airplane, the illusion projects the large plane to be much farther away than is its actual distance from the observer. This illusion also interacts with familiarity. If we expect to see a large plane and not a small plane, our initial per- ception will be biased toward the larger object farther away. Recall the role of familiar size in depth perception from Chapter 7. Additional cues must be put in place before we switch our perception to a smaller object mov- ing more closely by. Because of this, the perception of the speed of the airplane is affected, and the speed is judged to be less than it actually is. In this illusion, the perceived size and distance depend on an assumption of what the object is, that is, whether it is a small or large aircraft. This leads to the conclusion that familiarity with size affects our perception of speed of motion (Hershenson & Samuels, 1999). When planes experience technical trouble, the safety of the pilots and passengers is then most dependent on the pilot’s ability to perceive accurately. Consider a small plane being buffeted by powerful winds on a very dark night. Typically, a pilot can maintain the plane’s balance by instruments on his panel. These instruments measure the pitch and roll and help keep the wings straight despite the buffeting wind. However, if these instruments were to be lost due to technical problems, the safety of that plane would depend on the pilot’s ability to keep the plane from rolling too much during flight (Nooij & Groen, 2011). Although this perception is more vestibular (see Chapter 14) than visual (given it’s a dark night), proper training allows pilots to sense the correct orientation of the plane even without their instruments. Repetition of this task in simulators during training also allows pilots to be less affected by motion sickness. Indeed, spatial dis- orientation has been linked to many airplane accidents, and thus such training is necessary (Kallus, Tropper, & Boucsein, 2011). We count on our pilots to safely transport millions of peo- ple at high speeds across long distances every day. In most airports (e.g., Atlanta, Chicago, or Dallas), multiple planes may be taking off or landing at any single moment. Thus, studies that can reduce pilot error are critical. One such pilot error is illusory perceptions, including the size–dis- tance illusion discussed here. Luckily, there is an active research program in many countries that helps pilots fly more safely. Chapter 8: Movement and Action 251 CHAPTER SUMMARY 8.1 Explain the concept of a motion threshold and what variables influence that threshold. Motion is a change in position over time. We need to know if moving objects are approaching us or heading away from us. We must be able to perceive motion across three dimensions—directly ahead of us, to our sides, and above or below us. We must also be able to estimate the speeds of moving objects. Motion thresholds are a function of what part of the retina is seeing the motion. We have rather poor motion thresholds in the foveal regions of our retinae. But our motion thresholds are better at the periphery of our reti- nae. Apparent motion is the appearance of real motion from a sequence of still images. We see apparent motion every time we watch a movie. Induced motion means that one moving object may appear to make another object look like it is moving. Using fMRI, Larsen et al. showed that the areas of the primary visual cortex that were responsive to apparent motion were the same as those responsive to real motion. 8.2 Describe the anatomy of the visual system dedi- cated to motion perception. The retinae play an important role in early motion detec- tion. We know that some types of amacrine cells in the retina are sensitive to motion. In frogs, we find motion detectors, known as bug detectors, built right into the retinae. Reichardt detectors are hypothetical neurons that are sensitive to motion. They respond to activity in one receptive field at Time 1 and in another receptive field at Time 2. The corollary discharge theory states that the feedback we get from our eye muscles as our eyes track an object is important to the perception of motion. Thus, in addition to being sensitive to motion in receptive fields, we use our own eye movements to help us perceive motion. Saccades are the most common and rapid of eye movements. They are used to look from one object to another. Smooth-pursuit eye movements are the voluntary tracking eye movements that allow us to slowly move our eyes in concert with the motion of an object. MT (V5) is an area in the occipital cortex critical to motion perception. Newsome and his colleagues have shown that the mon- key’s MT is responsive to large-scale movement, that is, the movement of many objects together. Weigelt et al. showed that the human MT is active when we are attend- ing to motion in a moving display and even when we are just imagining movement. Akinetopsia (motion blindness) is a rare condition in which an individual is unable to detect motion despite intact visual perception of station- ary stimuli. It is caused by damage to area MT. A motion aftereffect is a motion-based visual illusion in which a sta- tionary object is seen as moving in the opposite direction of real or apparent motion just observed. 8.3 Discuss how we infer form from the perception of biological motion. A point-light walker display is one in which small lights are attached to the body of a person or another animal, which is then filmed as it moves in an otherwise completely dark environment. Even when we can see only the lights, we can see the form of the human just from the pattern of move- ment. Thus, we can infer form from motion perception. 8.4 Summarize the concept of an affordance. Affordances are the information in the visual world that specifies how that information can be used. One important aspect of visually guided action is optic flow. Optic flow is the relative motion of objects as an observer moves for- ward or backward in a scene. The gradient of flow is the difference in the perception of speed of objects moving past us in an optic flow display. The focus of expansion is the destination point in an optic flow display, from which point perceived motion derives. The lateral intraparietal (LIP) area is an area of the primate parietal cortex involved in the control of eye movements. The medial intraparietal (MIP) area is an area of the posterior parietal lobe involved in the planning and control of reaching movements of the arms. The anterior intraparietal (AIP) area is a region of the posterior parietal lobe involved in the act of grasping. REVIEW QUESTIONS 1. What is meant by the term motion threshold? How would you go about measuring the speed threshold at which we perceive motion? 2. What is real motion? What is apparent motion? Why would a visual system evolve to perceive appar- ent motion as real motion? What common brain Sensation and Perception252 mechanisms underlie the perception of both forms of motion? 3. What is a Reichardt detector? How might the visual system use Reichardt detectors to perceive motion? 4. What is the corollary discharge theory? How does it link eye movements to the perception of motion? 5. What is MT or V5? What area of the brain is it in? Describe one study using single-cell recording and one study using fMRI that support the role of this area in motion processing. 6. What are motion aftereffects? How are they experi- enced? What areas of the brain are involved in pro- ducing them? 7. What are point-light displays? How have they been used to examine the relation of form and motion per- ception? Describe an experiment that supports that role. 8. What is an affordance? How does the concept of an affordance serve to bridge the gap between percep- tion and action? 9. What is the lateral intraparietal area? Describe a study that links this area of the brain to its role in the relation of perception and action. 10. What is the rotating snakes illusion? Why are micro-saccades critical in perceiving this illusion? Why do we make micro-saccades? PONDER FURTHER 1. Anyone who has ever watched a cat play with a laser pointer or chase a lizard or mouse knows how sensitive to movement a cat’s visual system is. If you were design- ing a cat visual system from scratch (pun intended), what variables would be different from human motion perception? 2. In ISLE 8.8, there is a demonstration of corollary dis- charge. What is corollary discharge? And why is it important for a person to be able to perceive motion in the world? KEY TERMS Affordance, 243 Akinetopsia (motion blindness), 239 Anterior intraparietal (AIP) area, 246 Apparent motion, 230 Corollary discharge theory, 234 Correspondence problem (motion perception), 231 Focus of expansion, 244 Gradient of flow, 244 Induced motion, 231 Lateral intraparietal (LIP) area, 245 Medial intraparietal (MIP) area, 246 Motion, 228 Motion aftereffect, 240 Point-light walker display, 241 Real motion, 230 Reichardt detectors, 233 Chapter 8: Movement and Action 253 Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review, study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning. Learning Objectives Digital Resources 8.1 Explain the concept of a motion threshold and what variables influence that threshold. A Window on Reality: Perceiving Edited Moving Images Van Gogh’s “Starry Night” Optical Illusion 8.2 Describe the anatomy of the visual system dedicated to motion perception. The Woman From Munich Who Is Motion Blind 8.3 Discuss how we infer form from the perception of biological motion. 8.4 Summarize the concept of an affordance. Action’s Effect on Perception 9Visual Attention Victor Habbick Visions/Science Source LEARNING OBJECTIVES 9.1 Explain selective attention and divided attention, which help us understand our scanning of the world. 9.2 Illustrate how we move attention around a scene both by using eye movements and by using our peripheral vision. 9.3 Assess the nature of attentional capture and how it works in our visual system to have stimuli capture attention. 9.4 Examine attentional blink and change blindness as ways of helping us understand how attention changes over time. 9.5 Explain the differences between the orienting attention network and the executive attention network and their roles in directing visual attention. 9.6 Identify what factors lead to the development of visual attention as we grow up. INTRODUCTION In 2008, the classical violinist Joshua Bell brought his $3 million, 300-year-old Stradivarius into a busy Metro station in Washington, D.C. Bell, a renowned soloist who has toured the world playing the most complex pieces in the classical repertoire, plays to packed concert halls. A front-row seat can cost upward of $200. But on that Friday morning, Mr. Bell took out his violin near an escalator in a busy sta- tion and played his finest pieces for 45 minutes (Figure 9.1). Amazingly, few people stopped to listen, and only one person recognized him. Most people walked right by him, not realizing that a rock star of classical music was performing in a place where lesser known street musicians often play. What had happened? People were busy rushing to work, making sure they caught their trains on time, and presum- ably were more focused on their own lives than the beautiful music emanating across the train station. It is reasonable to guess that what had occurred was not a failure of music appreciation but rather an illustration of the critical nature of atten- tion. Because no one was attending to the musician in their midst, no one realized who he was (see Chabris & Simons, 2009). A major social concern in many countries is the scourge of texting while driving (Figure 9.2). People lead busy lives, and in many places, such as big cities, many people spend an inordinate amount of time stuck in traffic. The temptation to use that time to ISLE EXERCISES 9.1 The Stroop Effect 9.2 Scanning 9.3 Spatial Cuing 9.4 Inattentional Blindness Examples 9.5 Stimulus Salience 9.6 Feature vs. Conjunction Search 9.7 Change Blindness 9.8 RSVP 9.9 Attentional Blink and Repetition Blindness 9.10 Hemifield or Unilateral Neglect 9.11 Bálint’s Syndrome 9.12 Perceptual Bistability 9.13 Navigation in Blindsight 256 Sensation and Perception text to friends, family, or colleagues is strong. But research shows that texting (and talking on cell phones) is a major distraction and strongly impairs the ability to drive, result- ing in many accidents and many deaths. Drivers are slower to respond to dangers while texting and consequently get into more accidents. This outcome for cell phone use has been demonstrated both on the road and in the lab (Strayer, Watson, & Drews, 2011). But why? Why can’t we drive and text at the same time? Of course, part of the answer is that you are looking at a small screen and not at the road, so you can hardly make responses to objects on the road if you cannot see them. Recall from Chapter 3 that we have good acuity only in and near the fovea. So if the road is in the periphery, you cannot make it out clearly. But even if you are looking out over your cell phone so that you can direct your vision back and forth, you are still putting yourself and others at risk, because you have only a limited amount of visual attention, which, when directed to your phone, is not being directed to the potential hazards on the road. Thus, the reason why we cannot drive well and text at the same time revolves around the concept of attention. The question we concern ourselves with in this chapter is the nature of visual attention, its limits, how attention can be selected and divided, whether we are justified in thinking about atten- tion as unified or whether it must be considered multiple phenomena, and, finally, the neurological underpinnings of visual attention. Right up front, it is important to make a distinction among three psychological terms, which are all sometimes called attention in everyday life. We will see that each one has different psychological traits and different psychological functions. The terms are alertness, attention, and awareness. Alertness refers to a state of vigilance. When alert, we are awake, mindful, and scanning our surroundings. Alertness, however, implies that we are not attending to any particular stimulus; rather, we are waiting to find out what we should be paying attention to. For example, we want a security guard to be alert, looking out for any kind of danger at any time. A security guard asleep at his station is not providing much security. However, we do not want him focusing too much on one screen on his bank of monitors, because an intruder may be coming in through another entrance. We want our security guard to be alert but not directing his attention too much to any one source. Attention, the main focus of this chapter, is the allocation of our limited cognitive resources to one of many potential stimuli. This term implies selection. Thus, while visiting the zoo, you may see many individual animals of many species within one enclosure. But you attend to the cute baby giraffe following its mother around. At a restaurant, there may be many people and voices, but you attend to the person with whom you are dining. Moreover, once the security guard detects an intruder on one camera, we want him to attend to what the intruder is doing, as he notifies the police. Finally, awareness is active thought about something, which can be either physically present or just in our imagination. Thus, you are aware that the bright blue sky outside indicates a pleasant day. In general, we are aware of our surroundings whenever we are awake, but there are also bizarre disturbances of awareness. In this chapter, we discuss patients who have lost their awareness of their visual world, even though their eyes are still functioning (blindsight). FIGURE 9.1 Violinist Joshua Bell The violin soloist Joshua Bell played in a Washington, D.C., train station for 45 minutes, and most people passed by without attending to his artistry. © Z U M A P re ss In c. /A la m y FIGURE 9.2 Texting While Driving A major social concern is texting while driving. An important question is, Why can’t we safely text while driving?. © iS to ck ph ot o. co m /n yc sh oo te r 257 Chapter 9: Visual Attention Because issues of awareness are close to issues of consciousness, we devote some time to visual awareness at the end of the chapter. Reading a textbook on sensation and perception requires your complete attention. To understand psychophysical methods, neurophysiological techniques, and the termi- nology of brain anatomy, you must stay focused on your reading. If you let your atten- tion wander, you may find that you do not remember or understand the last sentence you read. The lead pipe transformed into a feckless pecan. Did you catch that last sen- tence, or has your attention already drifted to the person sitting near you in the library or the chattering voice on the television? Distraction is the negative side of attention. When we are attending to an object in the world, we gather much information about it, but it means that we are distracted from other stimuli in our world. In some cases, this distraction has interesting consequences. We will discuss inattentional blindness, which describes how we miss potentially important stimuli when our attention is directed elsewhere. And worry not—neither the lead pipe nor the pecan will be on the exam. Why is it that we cannot attend to more stimuli at any point in time? There is some evidence to suggest that the human brain just does not have the computing power to follow more than one possible incoming stream of information (Dewald, Sinnett, & Doumas, 2013). To compensate for this inability, our visual system (and cognitive systems) have evolved to pay attention to some stimuli but not others. In this chapter, we use the term attention to refer to several mechanisms that allow us to direct our perceptual processes to some stimuli but not others. That is, attention is not a single process but many processes (Chun, Golomb, & Turke-Browne, 2011). We start with the following definition of attention: a set of processes that allow us to select or focus on some stimuli. There are a few features of attention worth noting at the outset. First, our focus in this chapter is visual attention. But attention can be directed to any of our senses, such as auditory sources or somatosensory sources. Indeed, when we are enjoying a good meal, we can direct attention to our taste senses. Second, attention can be directed externally to perceptual features of the world, but we can also direct attention inter- nally to our thought processes or imaginal processes (Raymond, Shapiro, & Arnell, 1992). Thus, for example, when you think about the Mona Lisa, you may attend to your internal imagery, which conjures up a picture of the da Vinci painting. Can you see the landscape in the background of your mental image of the Mona Lisa? (After conjuring up your image, have a look at Figure 9.3.) In this chapter, as in this book, our focus is on external attention, that is, our attention to stimuli in the perceptual world. Attention can also be sustained or temporary. Sustained attention, which often requires sustained alertness as well, is required in many transportation-related jobs. An airline pilot may spend 10 or 11 hours in control of a plane carrying many hundreds of people. She must be alert but also ready to attend to any problem that surfaces either with the plane or with the weather. Changes have been made in recent years to give pilots time off so that their attention does not have to be engaged continuously for hours, but in some cases, such as a trans-Pacific flight, pilots must be able to deploy attention to their jobs for many hours straight. Temporary attention is an everyday phenomenon as well. We may pay attention to some stimuli for short time periods and then direct our attention to other stimuli. For example, you may look up from your reading to see if your dog is in need of food or a walk. For that moment, your attention is directed toward your dog rather than your book. You pat his head and then go back to the book. Now your attention is back on your reading. Attention can be overt. For example, when we are looking directly into someone’s eyes, we are telling that person that we are paying attention to him. But attention may be covert as well. We may be looking at our conversational partner but attending to news coming from a distant radio. Our first topic is selective attention, the how and why of focusing attention processes on one, and only one, visual input. Attention: a set of processes that allow us to select or focus on some stimuli FIGURE 9.3 Attention to Internal Sources How well were you able to focus on your internal representation of the Mona Lisa? Could you attend to the background in your visual image? Is it harder to attend to stimuli in imagery than it is when we can actually see the stimuli? M usée du Louvre 258 Sensation and Perception SELECTIVE ATTENTION 9.1 Explain selective attention and divided attention, which help us understand our scanning of the world. Consider the following situation. You are sitting in your family’s living room. Your father is watching a base- ball game on the television, your brother is playing a video game, your mother is talking on the telephone, and you are trying to study. Obviously, this is not an ideal scenario for you to concentrate on your studying. However, what you are trying to do is engage in selec- tive attention. You are trying to focus on your sensation and perception textbook and block out the competing distractions. This is the essence of selective attention, in which we direct our perceptual and cognitive resources to one stimulus among many potential stimuli. Selective attention can be defined as the processes of attention that allow us to focus on one source when many are present. Selective attention can be compared with divided attention. Divided attention occurs when we try to attend to competing sources of information (Figure 9.4). Thus, in the example, you may be trying to study and pay attention to the balls and strikes in the baseball game. To the extent that you are successful at this, your attention is divided. Thus, we can define divided attention as the process of attending to multiple sources of information. Sometimes we cannot seem to prevent stimuli from intruding on our cognitive processes. Some responses to stimuli seem to be automatic. Responses just jump out regardless of where we want to direct our attention. This situation is best described once we have an example under our belt. Try ISLE 9.1 before you continue. Your task will be to name colors. Let us see how you do. Make sure you have done the experiment in ISLE 9.1 before you continue reading. In this famous task, the Stroop task, you are asked to name the colors (e.g., red or blue) of things you see, some of which happen to be letters arranged in patterns called words. The problem is that it is not a square or circle of color you are naming but a word and not just a word but a word that is a color word. Sometimes the word and the color are the same (e.g., the word green displayed in the color green), but sometimes the color of the word is not the same as the color word (e.g., the word yellow displayed in the color blue). You are trying to name colors, so the word should not matter. You have directed your selective attention to the color and you should be able to ignore the word. Right? J. Ridley Stroop (1935) tried this task and found that people could not ignore the meaning of the words. Think of your own experience. It took longer to respond when the words and colors did not match, and you probably made more mistakes. You might even have gotten a bit frustrated. This study is the classic example of automa- ticity. Automaticity refers to those cognitive processes that do not require attention. They happen automatically. These automatic processes, both innate and learned, can extend what we can handle in the day, but they can also interfere because we have a harder time controlling them. Just try the Stroop task again. This chapter will focus on attention, but it is good to remember that some perceptual and cognitive processing can occur without attention. If you want to make a learned task automatic, it takes a lot of practice. If you are a musician or athlete, think of all of the drills you have done to learn basic skills, such as playing scales or dribbling a basketball. ISLE 9.1 The Stroop Effect FIGURE 9.4 Divided Attention Modern life requires us to constantly divide our attention, as this family seems to be doing. © S hu tt er st oc k. co m /Y ul ia Y as Pe Selective attention: the processes of attention that allow us to focus on one source when many are present Divided attention: the process of attending to multiple sources of information Automaticity: refers to those processes that do not require attention 259 Chapter 9: Visual Attention TEST YOUR KNOWLEDGE 1. Figure out what it might be like to not have selective attention. 2. Predict what advantages and disadvantages a person without selective attention would have. SPATIAL LIMITS OF ATTENTION 9.2 Illustrate how we move attention around a scene both by using eye movements and by using our peripheral vision. Attention and the Direction of Gaze in Space Typically, we attend to a particular location in space by directing our gaze to that location. Thus, when reading, you are directing your gaze at the words on the page, and when you glance up at the baseball game, you direct your gaze to the television screen. In almost all situations, what is being rep- resented in our foveae is what we are attending to. If our attention shifts, so does our gaze. When one changes one’s attention from the textbook to the baseball game, it is usually accompanied by a change in the direction of gaze. This move- ment of our eyes makes sense given that the best acuity is found on our foveae (recall Chapter 3). It is only in and near the fovea that we have enough detail to support many of the demands of attention. Even within a scene, we usually move our eyes around the scene, or scan the scene, to take it all in. For example, find the small blue flower in this photograph (Figure 9.5). It is small, so unless you are lucky, you need to move your gaze around the picture to find it. The main way you look around the image is to make saccades (recall Chapters 4 and 8). Researchers use sophisticated methods to track eye movements. They look for where the eyes pause between saccades, the fixations, because during saccades, the eyes move too fast to see clearly (e.g., Zanos, Mineault, Guitton, & Pack, 2016). In ISLE 9.2, you can simulate what an eye tracker might find as you searched for the flower in this image. You can try other images or load your own. An intriguing question we can ask, however, is whether it is possible to attend to locations in the world other than our gaze direction. That is, can I be looking at my textbook, but actually attending visually to the presence of the baseball slugger, Miguel Cabrera, at home plate? That is, is it possible to direct our attention to an object not on the fovea but in our periphery? The assumption in sports is that we are capable of directing our attention to places other than our direction of gaze. Think of the “no-look” pass in basketball. A basket- ball player may be running down the court looking to the left at a teammate ahead of him. Instead, he passes the ball to a player moving to his right. This may catch the opposing team off guard, and the no-look pass results in a score, thanks to what is called covert attention (Figure 9.6). Covert attention is when your visual attention does not line up with your direction of gaze. The opposing player assumes that atten- tion is directed toward the location at which the player with the ball is looking, but that ISLE 9.2 Scanning FIGURE 9.5 Scanning Try to find the small flower in this photograph. Covert attention: when your visual direction does not line up with your direction of gaze 260 Sensation and Perception player is actually attending to another location, allowing the recipient of the pass a better chance at making a bas- ket. These passes work because attention is usually asso- ciated with our direction of gaze, which lines up with the foveae. Covert attention is opposed to the more normal situation, overt attention, where our attention lines up with where we are looking. Michael Posner and his colleagues investigated this phenomenon in a series of interesting experiments in the late 1970s and early 1980s (see Posner, 1980). In this paradigm, a participant is directed to look at a central fixation light. In many experiments, the participant’s eyes are monitored to ensure that she is looking at the central fixation point. Trials will not start until the par- ticipant’s eyes are still and focused on the fixation point (Hughes & Zimba, 1985). The participant then sees an arrow pointing to either the right or left of the fixation mark. This arrow is a cue, which directs the participant’s attention in visual space (Figure 9.7). The task of the participant is to maintain fixation on the center point but direct her covert attention in the direction of the arrow. Then a light appears on either the same side or the side opposite the one to which the arrow pointed. When the light occurs on the same side as the cue, it is said to be a valid cue, but when it occurs opposite the cue, it is said to be an invalid cue. In Posner’s original experiment, the cues were 80% valid and 20% invalid. With an 80/20 mix, when the cue pointed to the left, the target occurred on the left 80% of the time (valid), whereas it occurred 20% on the right (invalid). When the cue pointed to the right, the target occurred on the right 80% of the time (valid), whereas it occurred 20% on the left (invalid). There were also neutral trials in which the cue did not point to a particular side and the target occurred 50% of the time on each side. In the neutral condition, the participant did not have a hint as to the side on which the target would occur. The task for the participant was to respond as quickly as possible when the target light came on. Posner’s question was, would covert attention to the cued location facilitate the response to that location, and would covert attention to the cued location inhibit the response in the opposite location? The investigators measured this by recording the reaction time to indicate that the participant had seen the target. In general, most young adults are quite fast at this task and can make their responses in about 250 milliseconds (ms; a quarter of a second) (see Figure 9.7 for an illustration of the procedure). Posner (1980) varied one other feature in this experiment. Sometimes the cue appeared at the exact same time as the target, but in other conditions the cue occurred at set intervals, up to 400 ms before the target. Thus, Posner could also look at the time course of any cuing effect. This variable was called stimulus onset asynchrony. Stimulus onset asynchrony refers to the difference in time between the occurrence of one stim- ulus and the occurrence of another, in this case, the cue and the target. If the stimu- lus onset asynchrony is zero, there is no difference between the occurrence of the cue and the target; that is, they occur at the same time. If the stimulus onset asynchrony is 200 ms, the target occurs 200 ms after the cue is presented. This variable turned out to predict the results. First, when there is zero stimulus onset asynchrony, there is no difference in reaction time between the valid and invalid signals. However, if the cue occurs 200 ms before the target, there is a distinct advantage, with faster reaction times for targets in the valid location than targets in the invalid location. This is also true at longer stimulus onset asynchronies. Neutral trials were of intermediate speed (Figure 9.8). Later work on this paradigm showed that relative to an appropriate neutral trial, most Overt attention: when your visual attention lines up with your direction of gaze, that is, your foveae Stimulus onset asynchrony: the difference in time between the occurrence of one stimulus and the occurrence of another, in this case, the cue and the target D av id L ia m K yl e/ N at io na l B as ke tb al l A ss oc ia tio n/ G et ty Im ag es FIGURE 9.6 Covert Attention Stephen Curry of the Golden State Warriors may be looking ahead, but he may also make a no-look pass to the right or the left. Thus, his attention is somewhere other than the direction of his gaze. 261 Chapter 9: Visual Attention of the effect was inhibition of invalid trials rather than facilitation of valid trials (Hughes & Zimba, 1985). This means that, relative to the neutral tri- als, the disadvantage for invalid trials was bigger than the advantage for valid trials. You can try a demonstration of this experiment on ISLE 9.3. What does this study tell us about attention? First, it tells us that we can devote attention covertly, that is, direct attention to a location in space that we are not looking at directly with our foveae. Thus, our intuition that we can be looking at our computer screen, but attending to the cute girl or boy at the other table in the library, is con- firmed. Similarly, it confirms that athletes can look one way but attend in the opposite direction, as in the aforementioned no-look pass. But we should not forget the loss of acuity in the periphery. Posner (1980) likened attention to a spotlight that we can shine on particular locations in space; that is, there is a spatial limit to attention. Visual attention has a size. In that spotlight, we can pro- cess what we see there in a more efficient manner than if the spotlight were not directed there. Much as a spotlight on a stage directs your attention to that location, your internal spotlight makes it eas- ier to process information in that location, while making other locations more difficult to pro- cess (Figure 9.9). On the basis of the previously described experiments, Posner also thought the spotlight could be directed away from the region a person is actually looking at. The spotlight met- aphor makes a good general model of selective attention in space, but we will see shortly how the spotlight is not always perfect, and we can make errors in processing even when we are attending to a particular location in space. An interesting extension of Posner’s paradigm was conducted by Egly, Driver, and Rafal (1994). As in Posner’s paradigm, a cue was given to indicate FIGURE 9.7 The Posner Cuing Paradigm The participant maintains fixation at the central point. Cues indicate whether the target will appear on the left or the right. In most trials, the cue is valid; that is, the target appears on that side. But in some trials, the cue is invalid; that is, the target appears on the opposite side of the cue. The experimenters measure the reaction time to indicate that the target is present. Start * * * * * * * * Time 1 Time 1 Tim e Time 2 Time 2 Time 3 X X X Time 3 Test probe Response Keys This cue is valid This cue is invalid Tim e Tim e 220 Invalid cues Neutral cues R e sp o n se t im e ( m se c) Valid cues 240 260 280 300 320 FIGURE 9.8 Results of Posner’s (1980) Experiment Valid cues result in faster reaction times to the target than do neutral (control) trials or invalid trials. This figure depicts the reaction time with a stimulus onset asynchrony greater than zero. ISLE 9.3 Spatial Cuing 262 Sensation and Perception where a target light would occur. However, the setup was a little different (Figure 9.10). A target could occur at any of four locations, as seen in Figure 9.10. A cue indi- cated where the target was likely to occur. The target then occurred most often in the cued location (valid cues) but occasionally in other locations. As in Posner’s paradigm, the target was responded to faster in the valid than the invalid locations. However, there was also an advantage for the invalid cue when the target appeared in the same object as the cued location. Notice that in Figure 9.10, locations A and B sit on the same object in the display, even though they are in different spatial locations. That is, we can see that Location B has a faster response time than Location C when it is Location A that is cued. This has been called the same-object advantage. Why? The attentional spotlight is focused on Location A, but it travels to B more easily than C because there is no boundary between B and A, but there is a boundary between A and C. Thus, response times to Area B also show some enhancement. It is also called the same-object advantage because a location of C is equal dis- tance from A as is B, but C is not physically connected to A as is B, and, thus, it would not benefit from the attentional focus. That is, Location C resides in a different object. Thus, Egly et al.’s (1994) experiment extends the con- cept of spatial attention. Areas in the object we are attend- ing to will benefit even if we are not directly attending to that area. Where might this finding apply in the real world? Think about two wrestlers competing with each other. One wrestler may be watching his opponent’s eyes for any clue as to what move he will make next. Attending to the wrestler’s eyes may allow that wrestler to more quickly sense his opponent’s hands moving in a particu- lar way than the hand movements of a spectator sitting nearby. The opponent’s hands are attached to his eyes, and thus it makes good sense if the attentional focus spreads easily to other areas along the same object. However, unless we are playing competitive sports, engaging in international espionage, or participating in Posner-paradigm experiments, we are usually not engaged in covert attention but in overt attention. This means that we are looking directly at what we are attending to, or when our attention is shifted, we direct our attention and our gaze to the new stimulus. It is this overt atten- tion, or rather failures of overt attention, that we examine next. Inattentional Blindness Inattentional blindness refers to a phenomenon in which people fail to perceive an object or event that is visible but not attended to. This refers to situations in which a well-above-threshold event or object is not seen because the person’s attention is directed elsewhere. In the 1963 movie Cleopatra, during one scene, an airplane is passing overhead. Given that the movie was supposed to take place in 48 BCE, this is clearly an error on the part of the filmmakers. However, most viewers will not notice FIGURE 9.9 The Spotlight Model of Attention The spotlight model argues that, much like a spotlight on a stage, attention focuses on one location in visual space and allows us to process information better there. © iS to ck ph ot o. co m /j gr ou p FIGURE 9.10 Experimental Setup From Egly et al. (1994) Participants look at the central fixation marker, but the target can occur in any one of four locations. The researchers found that cues to the left or right would have a faster reaction time even when the cue was directed to either the top or the bottom of the same object (same side). Cue Present cue Present targetCue off 374 msC + D A B C + D A B 324 ms 358 ms Inattentional blindness: a phenomenon in which people fail to perceive an object or event that is visible but not attended to 263 Chapter 9: Visual Attention the passing airplane, as it is irrelevant to the storyline and certainly not expected. Even if you watch it now, alerted to this anachronism, you will likely miss it. Movie directors who miss these goofs definitely benefit from inattentional blindness! Try an example. Read Figure 9.11. What is wrong in it? Did you find the extra the? Inattentional blind- ness is a counterintuitive finding. Most people would not predict it to occur and are quite surprised when they find that they are susceptible to it as well. Perhaps the most famous demonstration of inattentional blindness comes from a study conducted by Simons and Chabris (1999), which builds on an earlier example shown first by Neisser (1979). Before you continue reading, take a moment to view their video of this demonstration, which can be found at Chabris and Simons’s website (http://www.theinvisiblegorilla.com/videos.html) (see also ISLE 9.4). Like many of the illusions and phenomena we describe in this book, seeing is believing for this demon- stration. On Chabris and Simons’s website, the demonstration we are about to describe is the first video provided and it is also found on ISLE 9.4. Take the time now and view this video. It will definitely be worth it. In the experiment, Simons and Chabris (1999) asked participants to watch a video that shows three young college-age adults wearing white T-shirts, passing a basketball, and three young college-age adults wearing black T-shirts, passing another basketball. The six people are weaving among one another. The participants watching the video are instructed to count the number of times the team with white shirts passes the basketball among themselves. Less than a minute into the video, a person wearing a gorilla suit walks into the scene, contin- ues through, and then walks off camera. The “gorilla” is visible at the center of the video for about 5 seconds. Once the video is over, participants are asked how many passes they observed, and then they are asked if they saw the gorilla. Nearly half (46%) of participants failed to notice the “gorilla” walking across the field of view. Simons and Chabris describe the surprise that many participants experienced when they rewatched the video to determine that yes, the “gorilla” was visible and salient while they were attending to the basketball passes. In Simons and Chabris’s (1999) experiment, participants are focused on a challeng- ing perceptual task, counting the white team’s basketball passes while ignoring the black team’s basketball passes. It is actually quite a difficult task, and many participants were not able to count the number of passes correctly, regardless of whether they noticed the “gorilla.” Because of this, their attention was narrowly focused, which then caused them to ignore other stimuli to the extent that they did not even notice the very odd appearance of the “gorilla” (Figure 9.12). Inattentional blindness can be demonstrated under more traditional laboratory conditions as well. Mack and Rock (1998), for example, showed that participants were unable to see a perfectly visible stimulus when their attention was drawn to another aspect of the display. Participants main- tained fixation on a central point (Figure 9.13). They then saw a large cross somewhere in the parafoveal region, that is, near the fovea. The participants’ task was to determine if the horizontal or vertical lines of the cross were larger. In some trials, but not all, the fixation point transformed into a diamond shape at the exact instant the cross appeared. Remember that the term fixation point refers to the point at which participants’ eyes were supposed to be focused. In trials in which the fixation point changed, participants were asked if they had seen anything change when the cross appeared on the screen. A majority of the participants did ISLE 9.4 Inattentional Blindness Examples FIGURE 9.11 Read This Sentence Read this sentence. Did you find the error? Read the text for more information. Flowers bloom in the the spring. FIGURE 9.12 Inattentional Blindness (Simons & Chabris, 1999) When viewed as a static photograph, it is hard to imagine that participants do not see the gorilla. 264 Sensation and Perception not report seeing the diamond, even though it was being presented directly onto the foveae. Thus, even when we are looking at some- thing directly, we may not notice a change when our attention is directed elsewhere. Is it important that people are unable to detect crosses changing to diamonds in simple displays or even fail to see a person dressed up as a gorilla in a contrived video? The question, of course, is, does inattentional blindness occur in real-world situations? The answer is yes. In a stunning demonstration of this phenomenon, Drew, Võ, and Wolfe (2013) used expert radiologists as their test participants. The stimuli were computed tomographic (CT) scans taken to determine if patients had dangerous nodules on their lungs. The radiologists in the study had many years of experience analyzing CT scans of lungs, looking for these lung nodules. Embedded in one of the CT scans was a small image of a gorilla, about the same size a radiologist would expect a lung nodule to be (Figure 9.14). Eighty-three percent of the radiolo- gists failed to detect the image of the gorilla, even though eye-tracking software showed that many of them were looking right at the image. The radiologists were better than the nonexperts. Not one of the nonradiologists in the study noticed the gorilla. Nonetheless, even experts, operating within their sphere of expertise, show inattentional blindness. TEST YOUR KNOWLEDGE 1. Classify the different ways that we can direct our attention around the world. 2. Evaluate what inattentional blindness tells us about how we use attention. STIMULUS FEATURES THAT DRAW ATTENTION 9.3 Assess the nature of attentional capture and how it works in our visual system to have stimuli capture attention. We have discussed a failure of attention, inattentional blindness. Part of the explana- tion for inattentional blindness is that our attention is directed elsewhere. We are not attending to the stimulus an experimenter thinks is relevant, just something else. In this section, we will discuss some of the features of stimuli that draw our attention. FIGURE 9.13 Inattentional Blindness: Stimuli From Mack and Rock (1998) In critical trials, the fixation point changed from a cross to a diamond. In many critical trials, participants failed to notice the change. Noncritical trial Critical trial Fixation point Cross with arms of different lengths Fixation 1,500 msec Stimulus 200 msec Time Time Mask 500 msec Fixation 1,500 msec Stimulus 200 msec Mask 500 msec Critical stimulus 265 Chapter 9: Visual Attention Stimulus Salience You are driving down the road, and one of those new elec- tronic billboards flashes to the next advertisement. You look up at the billboard; that is, it is taking your atten- tion. The billboard captured your attention using what is called stimulus salience. Stimulus salience refers to the features of objects in the environment that attract our attention. Salience can be any number of features—bright colors, fast movement, personal relevance, or, in the non- visual domain, a loud or distinctive sound or smell. From the auditory domain, think about nodding off during a boring lecture in class, and then the professor calls your name. Hearing your name is a salient stimulus, and you are immediately paying attention—why has she called on me? Another example might be seeing a person streak- ing by without clothes in your university library. Because this image (positive or negative) is novel, surprising, and potentially important, your attention is immediately diverted from your calculus homework to the sight of the person streaking by. In general, stimuli that are novel or unexpected will act to divert our attention to them. The process by which a stimulus causes us to shift atten- tion is called attentional capture (Anderson & Yantis, 2013). In Figure 9.15, we illustrate attentional capture. Here we have a novel situation—a dog wearing reading glasses. Your attention is captured by the dog’s glasses because this aspect of the photograph is unusual. Most dogs indeed prefer chasing cats and ducks to reading the newspaper. You can also try the images in ISLE 9.5, which uses the simulation of eye tracking from ISLE 9.2. Look around the images and see what features of the image grab your attention, that is, what parts of the image have stimulus salience. They should be where you spend the most time with your eyes. Brian Anderson and Steven Yantis have explored the notion of attentional capture. They found that one feature that captures our attention is a stimulus that has been previ- ously associated with reward. Rewards are generally considered positive experiences, and we seek them out. Thus, a stimulus associated with a reward ought to attract our attention. Think about those “guns” or launchers that shoot T-shirts at sporting events, which are now nearly ubiquitous at American sporting events. Fans get free T-shirts if they are lucky enough to have one shot in their direction. Because these T-shirt launchers are associated with rewards, they may capture our attention if we see them in a different context later. Anderson, Laurent, and Yantis (2013) conducted an experiment in which they paired particular colored stimuli with specific amounts of monetary reward. The participants learned to associate these stimuli with their linked rewards. Later, participants engaged in a different task using different objects in different colors. However, Anderson et al. peri- odically presented a nonrelevant stimulus in a color associated with the highest reward in the earlier task. They found that when participants saw this reward-associated object, it captured their attention and slowed their performance on the intended task. Thus, stimuli that have value can also capture our attention. FIGURE 9.14 Inattentional Blindness in the Real World In this intriguing but frightening study by Drew et al. (2013), most radiologists failed to notice a small image of a gorilla on a CT scan they were asked to inspect for evidence of lung disease. Slice 1 Slice 2 Slice 3 Slice 4 Slice 5 50% 75% 100% Gorilla Opacity 75% 50% Stimulus salience: refers to the features of objects in the environment that attract our attention Attentional capture: the process whereby a salient stimulus causes us to shift attention to that stimulus ISLE 9.5 Stimulus Salience FIGURE 9.15 Attentional Capture A dog wearing reading glasses may attract your attention more so than a less unusual image. © iStockphoto.com /Fly_dragonfly 266 Sensation and Perception Semantic meaning may also capture our attention in a visual scene, if that meaning is presented in the fovea. Võ and Henderson (2011) showed participants pictures such as the ones seen in Figure 9.16. In one version of the picture, participants saw a kitchen scene with a pot on the stove. In a second version of the scene, participants saw a kitchen scene with a computer printer on the stove. In an earlier experiment, participants could move their eyes freely around the scene (Võ & Henderson, 2009). In this 2009 study, they found that participants’ attention was drawn to the printer and that participants spent more time looking at it than when the object was an expected pot. However, in the later study (Võ & Henderson, 2011), participants’ ability to move their eyes around the scene was lim- ited by the technology used to present the stimulus. In this study, atten- tional capture did not work at the periphery. When participants could see the printer only at the periphery, it did not attract attention in the form of eye movements. This was true even when the printer was displayed suspended in space above the stove. Thus, attentional capture works best when the stimulus evoking capture is processed by the fovea. Consider the situation described by Võ and Henderson (2011). We are looking at a common kitchen scene but fail to notice the printer miracu- lously defying gravity and suspended in the air above the stove. In real life, of course, we would be moving our eyes about the scene and would pick up this anomaly with our foveal vision. But nonetheless, this failure to detect unusual objects constitutes a failure of attention. It should be adap- tive to notice unusual circumstances, as such circumstances may present a threat, especially when we see them only peripherally. With stimulus salience, we have been discussing moving our eyes around to different parts of a scene. Let us look further into looking around a scene for some- thing we want to find. Visual Search One of the most important attentional tasks in vision is visual search, that is, looking for and finding one object amid a background of visual distrac- tion. Think about those security personnel at the airport. They must examine X-ray screens looking for banned items, such as weapons, among the vast array of objects that people bring onto airplanes. In this case, screeners must distinguish between rifles and golf clubs and between knives and flutes. One can think of few visual searches more important than those airport screen- ers do countless times every day at airports. But passengers must engage in visual search, too, when they are at the airport. When you arrive at your destination, you must be able to distinguish the face of your sister among a sea of other faces also waiting for passengers. In a more peaceful setting, away from the airport, a birdwatcher may be looking for the rare eider duck among a flock of mallard ducks. Finally, young children may spend endless time engaged in visual search, trying to see “where’s Waldo.” In Figure 9.17, you can engage in visual search to look for the dead tree and the satellite dish. How long did it take you to find these objects? We can define visual search as looking for a specific target among distracting objects. In the laboratory, we can control many aspects of visual search. We can control the size, shape, color, and location of the search item as well as the size, shape, color, location, and number of distracting items. We can then measure how long it takes a participant to find the target under these different circumstances. Consider Figure 9.18a. This figure demonstrates a FIGURE 9.16 Attentional Capture Participants’ attention was drawn to the printer, and participants spent more time looking at it than when the object was an expected pot. However, when it was presented at the periphery, participants failed to notice the printer miraculously defying gravity and suspended in the air above the stove. 267 Chapter 9: Visual Attention Visual search: looking for a specific target among distracting objects Feature search: the search for a target in which the target is specified by a single feature Conjunction search: the search for a target in which the target is specified by a combination of features Feature integration theory: a theory stipulating that some features can be processed in parallel and quickly prior to using attentional resources, whereas other visual characteristics require us to use attention and are done serially and therefore less quickly ISLE 9.6 Feature vs. Conjunction Search FIGURE 9.17 Visual Search Can you find the dead tree and the satellite dish? feature search. You are looking for the vertical orange bar amid a mix of different stimuli. Feature searches such as these are very easy. In Figure 9.18b, we see a conjunction search. Here you are still looking for the vertical orange bar, but now the distractors include both horizontal orange bars and vertical blue bars. You are searching for a conjunction between two features—vertical and blue. Because of this, the task gets more difficult with a greater number of distractors. You can vary the set size (number of distractors) and a host of other factors in ISLE 9.6. The feature search in Figure 9.18a is fast, and the number of distractors is irrelevant. We can pick out the orange bar quickly no matter how many blue bars are shown in the figure. This kind of sudden visual search is called “pop-out,” because the target seems to jump out at us from the display. When we are given a feature search, we seem to be able to do the search in parallel (Wolfe, 2012). The conjunction search in Figure 9.18b is different. Here we are looking for the vertical orange bar among some stimuli that are also vertical and some stimuli that are also orange. In this case, we tend not to have a pop-out effect. Instead, we must engage in a serial search, that is, looking at every object until we find the one we are looking for. Because of the conjunctions, we can think of these searches as being less efficient. Feature Integration Theory One of the prominent theories advanced to account for visual attention is feature inte- gration theory (Treisman & Gelade, 1980). This theory stipulates that some features can be processed in parallel and quickly prior to using attentional resources. Other visual characteristics require us to use attention and are done serially and therefore less quickly. In this view, there are some characteristics that simply pop out at us, such as the gross mispelling of this wird. Most of you instantly noticed that word was misspelled, but you might have missed that misspelling was also misspelled. Treisman and Gelade (1980) argued that many perceptual characteristics simply pop out, but only those that do not require attention. Consider the display in Figure 9.18b and 9.18c. The pattern shows that conjunctions and configurations require attention. We must actively search in time to determine the mismatched figure. In contrast, the searches that do not require the finding of a conjunction or configuration tend to pop out at us. 268 Sensation and Perception (a) Feature search Find Set size Find Find or (b) Conjunction search (c) Spatial con�guration search Correct target present Correct target absent ∼0 ms/item R e a ct io n t im e ( m s) 10∼3 0 m s/ite m 5∼15 m s/item 40 ∼6 0 m s/i te m 20∼ 30 m s/it em FIGURE 9.18 Visual Search Tasks In a feature search (a), the participant must identify the object that differs along one dimension (here, color). In a conjunction search (b), the participant must find the unique object among two sets of objects. Finally, in a spatial configuration search (c), the participant must look for a particular shape among numerous related shapes. The feature search takes place in parallel and therefore takes no longer among several distractors than many distractors. However, the conjunction search and the spatial configuration search are done in a serial self-terminating manner, which requires more time the more distractors are present. TEST YOUR KNOWLEDGE 1. Contrast the role of visual salience in capturing attention with how we direct attention with eye movements and covert attention. 2. Collect data from feature and conjunction searches in ISLE 9.6, and examine how the reaction times change, or not, as the number of distractors increases. Examine what it is about the feature search that leads those targets to pop out, and predict what other visual features might also lead to pop-out. ATTENTION OVER TIME 9.4 Examine attentional blink and change blindness as ways of helping us understand how attention changes over time. Change Blindness Imagine this situation: You are talking to a person at a party, and a group of people separate you for a moment. What if the person you had been talking to leaves and you 269 Chapter 9: Visual Attention end up talking to another per- son? Would you notice? Change blindness is the difficulty we experience in detecting differ- ences between two visual stimuli that are identical except for one or more changes to the image. Change blindness is a very coun- terintuitive effect. We usually think we will notice obvious changes when they occur, but change blindness suggests that we often fail to do so. What is inter- esting about change blindness is that even when we have engaged our attention and are directing our visual search in a very conscious manner, we may still fail to see the changes in an image. Consider Figure 9.19. Can you find the change? You probably can, but it does not immediately pop out for most people viewing these images. In a classic demonstration of change blindness, Simons and Levin (1997) conducted a field experiment in the city of Ithaca, New York. Participants were chosen at random on the Cornell University campus. In the experiment, an experimenter posing as a tourist asked a person on campus for directions. When the participant started giving directions, two other experimenters, posing as construction workers, passed by carrying a detached door. The construction workers passed between the experimenter who had asked for directions and the participant. At this point, Simons and Levin had the first experimenter hide behind the door and exit the scene, while a second experimenter, the same gender and age as the first experimenter, replaced the first. The question was whether the participant would notice that the person to whom he was giving directions had changed. It seems obvious—if you were just talking to one person, and he was replaced by another person, of course you would notice the difference. However, this obvious answer turned out to be wrong most of the time. Simons and Levin (1997) asked both professional researchers and undergraduates to predict the outcome of this experiment before the results were made public. Both pro- fessionals and students predicted overwhelmingly that almost all people would notice the change. After all, the participant finds himself talking to a different person. When the predictors were asked if they themselves would notice the difference, 95% of people answered in the affirmative. But this is not what Simons and Levin found in the field. In the study, nearly 50% of the participants failed to notice the change from one person to another. Indeed, Simons and Levin have made available video of participants con- tinuing to explain and point without noticing that they are giving directions to a new person. Check out a video the researchers have shared online at ISLE 9.7. Let us repeat: Nearly half of the participants failed to notice that the people they were talking to had changed from before the workers passed by to after the workers had passed by. Do these findings generalize to events in the real world? One such situation might be detecting errors or “goofs” in movies. Hollywood has professional staffers who, during postproduction, go through films to detect inconsistencies. However, invariably, some of these inconsistencies get through and make it into the movie. Fans and aficionados look for these errors and catalog them on websites, but most of us casual moviegoers never notice them. Of course, nowadays, it is rather easy to look up these goofs online before we go to movies and then annoy our friends by pointing them out, but most of us watch these movies originally without ever noticing these errors. For example, in the movie Iron Man 3 (2013), the characters of Tony Stark and Colonel Rhodes are talking together via cell phone to the vice president. The vice president’s cell phone is different from one shot to the next. Most of us, including the editing staff of the movie, failed Change blindness: the difficulty we experience in detecting differences between two visual stimuli that are identical except for one or more changes to the image ISLE 9.7 Change Blindness FIGURE 9.19 Change Blindness Can you detect the difference between the photographs? If you cannot find the difference and give up looking, go to the end of the Chapter Summary for the answer. 270 Sensation and Perception to notice this when seeing the movie. Consistent with the concept of change blindness, such errors abound in movies, but we almost always fail to notice them. In a more typical experiment on change blindness, Rensink, O’Regan, and Clark (1997) showed photographs of a variety of scenes, one image at a time. After a set period of time the photograph would disappear and be replaced by a nearly identical image. In the new photo, one aspect of the image would be different, as in Figure 9.19. The partic- ipants’ task was to identify what had changed. If they could not immediately detect the change, then participants were given a set amount of time to flip between the two images and detect the change. Here, too, some participants never found the change between one image and the other. You can try a change blindness similar to this study in ISLE 9.7. Thus, particularly if the change from one photograph to the other is not to an important aspect of the meaning of the photograph, we often fail to notice the change. Moreover, the Good Samaritan giving directions is not all that concerned with what the lost person looks like. She is simply doing the right thing by giving directions and may not really be paying attention to the person making the request, especially if the directions are complicated. She may be focusing on a mental map of the area. Thus, we can think of change blindness as a deficit in attention. One can ask what is going on at the neural level when we fail to notice these changes. After all, we are looking at different images but perceiving them as the same. Can we see different patterns of brain activity when we are blind to changes and when we actually recognize the changes? This topic was investigated in an experiment by Busch (2013). Busch showed par- ticipants a photograph of a real-world object and then showed a second photograph later in which one of the objects in the photograph had been switched, similar to Figure 9.19. As in all other work on change blindness, many participants failed to recognize the change between the photographs. In Busch’s study, the participants did a final recognition task in which they had to identify objects as being part of the study task or new to the study. Some of these were the original object, some were changed objects, and some were entirely novel. This part of the study was done while participants were undergoing electroencephalographic (EEG) recording. Busch found that the EEG patterns were different for old objects and new objects, but also that the EEG patterns were similar for items whose change had been detected and those whose change had not been detected. Thus, some aspect of the change was being encoded, as the changes led to neural changes, even if they did not lead to the participants’ perceiving the pictures as “different.” Beck, Muggleton, Walsh, and Lavie (2006) used transcranial magnetic stimulation to examine change blindness. They inhibited the right parietal lobes of some partici pants and the left parietal lobes of other participants while those participants were doing a change blindness task, similar to the ones just described. They found that interrupting the function of the right parietal lobe increased change blindness in the sense that participants were less able to detect the differences between photographs. On the other hand, interrupting function of the left parietal lobe had no effect on change blindness. This result suggests that the right parietal lobe may be critical in this form of attention. As we will see later, the right parietal lobe is critical in many aspects of attention in addition to change blindness. Attentional Blink and Rapid Serial Visual Presentation Consider, again, the overworked airport security agent (Figure 9.20). She is tasked with monitoring the contents of bags moving through the X-ray detector at the airport. But FIGURE 9.20 The Airport Security Agent Consider again the overworked airport security agent. Dozens of bags pass through this machine every minute. This worker is attending to one location but looking for changes in that space across time. Jo hn M oo re /G et ty Im ag es N ew s/ G et ty Im ag es 271 Chapter 9: Visual Attention dozens of bags pass through this machine every minute. Essentially, from the perspective of the agent, she is attending to only a small area of space, but she must maintain this attention to particular objects across time. The agent never knows exactly when the form of a handgun may attempt to slip past. Also along law enforcement lines, think of the role of a SWAT team sniper. She must attend, through the scope of a rifle, to the people passing before her. The job here is to locate in time when the terrorist, rather than the innocent hostage, is in the sight of the rifle. For a less violent example, think of a line worker at a factory. He may have to inspect each set of spark plugs that goes by for any design flaws. As in airport security, this worker is attending to one location but looking for changes in the properties of what is in that space across time. Attention over time rather than space has been studied with a paradigm called the rapid serial visual presentation (RSVP) paradigm. In RSVP, a series of stimuli appear rapidly in time at the same point in visual space. Indeed, the stimuli may appear as fast as 10 items per second. The stimuli are usually letters or pho- tographs (Figure 9.21). The task of the partici- pant is to determine when a particular stimulus appears and to press a button or key as fast as possible after that stimulus occurs. Thus, the participant might be following a series of letters flashed 10 per second. The participant’s task is to press a response button every time the letter S occurs. In another version of the RSVP task, photographs might be flashing by at 5 photographs per second. The participant must hit the response button every time a photograph of a guitar player occurs. To try this task yourself, go to ISLE 9.8. The RSVP paradigm allows the researcher to ask many questions about what enhances the attentional focus on the stimulus to be responded to, and the paradigm can also allow investigation of what factors can distract our attention. For example, Zivony and Lamy (2014) asked participants to respond when a stimulus was a particular color. The stimuli to be judged were all presented at the point of fixation so that the images would be main- tained on the fovea. However, they also presented various distractors at the periphery to determine if these distractors would affect attention and interfere with performance. In their study, they found that distractors at the periphery that were the same color as the intended target attracted attention and reduced accurate performance on the primary task. Think of it this way. You are driving down the road looking for signs that indicate dangerous curves. Well, these signs are yellow. You are more likely to be distracted by other yellow signs like a road closed or stop ahead sign rather than a white speed limit sign. In another study, Irons and Remington (2013) asked participants to track two simul- taneous RSVP streams, one on each side of fixation. The participants’ task was to respond to a conjunction of color and location. Thus, they might be asked to respond to a red object on the left but a green object on the right. This made the task quite a bit harder, and therefore, overall errors were relatively high and reaction time was down. But they also found that red objects presented on the right captured attention and slowed participants’ responses even further. This tendency of red to capture attention could explain the utility of making the stop sign red. ISLE 9.8 RSVP Rapid serial visual presentation (RSVP) paradigm: a method of studying attention in which a series of stimuli appear rapidly in time at the same point in visual space FIGURE 9.21 The Rapid Serial Visual Presentation (RSVP) Paradigm In this study, a series of stimuli, here letters, appear rapidly in time at the same point in visual space. The participant presses a button or key as fast as possible after a particular stimulus occurs. Letter on, 15 ms Letter off, 85 ms Betw een 7 and 15 random letters before T1. Another 8 random letters after T1. Here, T2 is in position 2. T1, a white letter Position 1 (after T1) Tim e Position 2 (after T1)T2, the letter X + D M T1 L H X 272 Sensation and Perception Let us return to our security personnel looking at a stream of stimuli going before them on their screen. The question here is what finding one gun does to the ability to find the second gun right after the first gun. If we return to the simpler version of RSVP in which one must track only one stream of letters, we can examine this question by varying the time between the presentation of one target and a second target. Thus, if we are following letters and responding to S, there might be a lag of 4 letters between one S and the next, or there might be a lag of 10 letters between one S and the next target (typically another letter). This variation on RSVP allows us to ask our question about if detecting the first target reduces our ability to detect the second target. It turns out that it does, if the second stimulus occurs within 500 ms (half a second) of the first stimulus. This phenomenon is called attentional blink, which refers to the tendency to respond less reliably to the second target in an RSVP task. Indeed, in some cases, the participant may not report having seen the second stimulus at all. In a typical study on attentional blink, the first and the second targets may be different. Thus, the partic- ipant knows to look for an S and then after the S to look for a K. One can then vary the lag between the S and the K and determine when and how the participant responds to the K. To demonstrate that attentional blink is an issue of attention, we merely have to change the instructions and ask participants to look only for the K. Given the same stream of stimuli as with the S-K instructions, K-only instructions yield faster and more accurate identifications of the K. The standard explanation of attentional blink focuses on an inhibition mechanism that dampens responses to other targets while processing the first target. When that target is detected, there is a finite amount of time necessary to “reboot” the attentional mechanism for the next target. Variables that slow this disengagement of inhibition can therefore strengthen the blink. For example, simply inducing fatigue in participants increases the intensity and length of the blink period (Kawahara & Sato, 2013). In contrast, there is some evidence to suggest that experienced video game players may show reduced attentional blink because of their training in games that resemble RSVP tasks (Kristjánsson, 2013). Finally, when the second target is the same stimulus as the first target (e.g., S-S) and occurs right after the first target, some participants will fail to see it at all. This phenomenon is called repetition blindness. You can see illustrations of both repetition blindness and atten- tional blink in ISLE 9.9. TEST YOUR KNOWLEDGE 1. Identify the role that the attentional blink might have in increasing the problems associated with using devices while driving. 2. Predict what elements of a scene might be subject to change blindness, and evaluate why we might be subject to this entertaining, irritating phenomenon. THE ANATOMY AND PHYSIOLOGY OF ATTENTION 9.5 Explain the differences between the orienting attention network and the executive attention network and their roles in directing visual attention. There are numerous questions that can be asked concerning the interaction of atten- tional processes in the brain. For example, we could take the standard brain localization approach and ask the following: What areas of the brain may be involved in the creation of attentional processes? Are there different brain regions for different forms of attention? Attentional blink: the tendency to respond less reliably or not at all to the second appearance of a target in an RSVP task when the second target occurs within 500 ms of the first target Repetition blindness: the failure to detect the second target in an RSVP task when the second target is identical to the first one; like attentional blink, it occurs when the second target is presented 500 ms or less after the first target ISLE 9.9 Attentional Blink and Repetition Blindness 273 Chapter 9: Visual Attention This standard approach of correlating anatomy and func- tion has yielded numerous interesting findings, which we discuss here. But we can also ask other questions, such as, when we are attending to one stimulus, how does that change the neural processing of that stimulus? Put another way, does attention change the way sensory areas of the brain respond to a stimulus, or does attention occur after such processing has occurred? For example, if I am focusing my attention on the pitcher in a baseball game, how does that change the neural processing of the visual image of the pitcher releasing the ball relative to a situation in which I was focusing my attention on the batter? We can also exam- ine what happens to the processing of nonattended visual stimuli. So, if I am focusing on the pitcher, does my percep- tual processing of the runner on first base change? We start with the basic anatomy of attention. The Orienting Attention Network Much of what we have presented as visual attention in this chapter is a function of a neural system called the orienting attention network (also called the dorsal attention network), which allows us to engage in visual search and direct our visual attention to different locations in visual space. The orienting attention network is based in circuits in the parietal lobe (Posner & Rothbart, 2013). Damage to this network can cause a number of neuropsychological conditions, most famously unilateral neglect (hemifield neglect), which we consider later in the chapter. Anatomical regions associated with the orienting attention network can be seen in Figure 9.22. The Executive Attention Network Beyond the orienting attention network, we need some neural processes to choose the stimulus to attend to. This function, among others, seems to be the domain of the exec- utive attention network. The executive attention network focuses on attention by the inhibition of habitual responses and acts as the top-down control of attention (Posner & Rothbart, 2013). This network allows us to inhibit auditory stimuli so that we can concentrate on visual stimuli, or it can allow us to inhibit visual stimuli so that we can concentrate on auditory stimuli. It also operates on attention directed at memory and higher order cognition. For example, in the context of texting and driving, this is the network that directs our attention to the small screens on our cell phones when we should be attending to the road. It is also the system that must be engaged when we attend to the color of a word (e.g., green written in red print) rather than its mean- ing (the famous Stroop effect discussed earlier in this chapter; review ISLE 9.1 for an example of this effect). The executive attention network is the product of processes in the prefrontal lobe. In models of memory, the executive attention network is sometimes called the “central executive,” because it comprises the brain regions by which attention is directed to the desired stimulus. Tamber-Rosenau, Esterman, Chiu, and Yantis (2011) conducted a functional mag- netic resonance imaging (fMRI) study with human participants to examine how these two attentional networks allocate attention in a visual attention task. Participants watched an RSVP display, consisting of two streams of letters, one to the left of fixation and one to the right of fixation. The researchers also used a distractor RSVP display above and below fixation to increase the difficulty of the task and to ensure that the participants would need their attentional networks. The above and below streams were Orienting attention network (dorsal attention network): a neural system, located primarily in the parietal lobe, that allows us to engage in visual search and direct our visual attention to different locations in visual space Executive attention network: a system that focuses on attention by the inhibition of habitual responses and acts as the top-down control of attention; found in the frontal lobe FIGURE 9.22 Attention Networks in the Brain EAN OAN 274 Sensation and Perception not to be attended to but rather provided background distraction. However, partic- ipants had to monitor both the left and right streams. When the letter L occurred, participants were to shift their attention from the left stream to the right stream of letters. When the letter L occurred again, participants shifted their attention back to the left stream of letters. The question Tamber-Rosenau et al. were interested in was what areas of the brain would show activity during these attentional shifts. Their results con- firmed involvement of both the executive attention network and the orienting attention network. They found that areas of the brain in the prefrontal lobe and in the parietal lobe were active during these attentional shifts. In particular, within the orienting atten- tion network, the medial parietal lobule was active during shifts of attention. Within the executive attention network, there was activity in the superior frontal sulcus/gyrus (Figure 9.23). The changes in these areas reflect these two attentional networks at work. We discuss shortly a similar study that shows how visual areas of the brain change in response to attentional changes. How Attention Affects the Visual Brain Think about the Posner paradigm described earlier (Posner, 1980). In this task, people are asked to direct their attention to a spatial location other than the location on which their eyes are fixated. To engage in this covert attention, participants must inhibit their tendency to want to move their eyes to the location of the to-be-attended stimuli. They must also inhibit the tendency to remain focused on the point of fixation. Thus, it is likely that this task draws on the executive attention network. But how does this net- work exert its attentional effect on the task? Attention is useful only if it can alter the efficiency of other cognitive processes. Thus, neutrally, attention must act on other areas of the brain that are otherwise engaged in the visual task at hand—detecting the target and making the response. So, attention must show up as both a “command” from an attention network and a change in perceptual processing in visual areas of the brain. FIGURE 9.23 Results From Tamber- Rosenau et al.’s (2011) Study These fMRI illustrations show activation in both the prefrontal lobe and the parietal lobe during attentional shifts. Within the orienting attention network, the medial parietal lobule was active during shifts of attention. Within the executive attention network, there was activity in the superior frontal sulcus/gyrus. 275 Chapter 9: Visual Attention This aspect of attention was illustrated elegantly in an experiment by Moran and Desimone (1985). Moran and Desimone (1985) conducted a sin- gle-cell recording experiment with rhesus monkeys. Monkeys were trained to keep their eyes fixated on a central point and then, as in the Posner spatial cuing paradigm, to attend to objects in one of two locations they were not directly looking at (Figure 9.24). The question the researchers were interested in was how attention would affect the physiological response in cells that were responding to visual characteristics, such as color and orientation. For this reason, they recorded cells in area V4 of the brain, which, as you will recall, is sensitive to both color and orientation. One of the presented objects was considered the effective stimulus because it elicited a large response from the cell in V4 being studied when presented alone in the receptive field. The second stimulus was considered ineffective because it did not elicit a strong response from the same cell in V4 when pre- sented alone in the receptive field. Thus, when both stimuli are presented to the visual field, without an attentional manipulation, one would expect a large response driven by the effective stimulus. However, in the study, the monkeys were directed to attend to either the effective or the ineffective stimulus. Note, of course, that there are presumably other cells in V4 that respond to the ineffective stimu- lus. But the cell being recorded from does not. When attention is directed to the effective stimulus, the V4 cell responds strongly, in fact, more strongly than when attention is not directed there. However, when attention is directed to the ineffective stimulus, the cell’s response decreased, despite the fact that the effective stimulus was in the receptive field as well. Because the effective stimulus was not being attended to, however, the response of the V4 cell decreased. Thus, this experiment showed that, at a physiological level, attention can affect the processing of visual information and at a fairly early stage of visual processing. Here, attention acts to decrease the activity of a cell when the effective stimulus for that cell is not being attended to. We can see similar patterns of activity in the human brain. Using fMRI, Chiu and Yantis (2009) examined shifts in activity within the occipital lobe as a function of attention. In their study, participants fixated on a central marker and then rapidly shifted attention from stimuli on the left side of fixation and from stimuli on the right side of fixation. The stimuli were either letters or digits. When the letter R occurred, it meant to direct atten- tion to the right and keep it there, until participants saw the letter L, and then they were supposed to direct their attention to the left. When attention was directed to the left, there was more activity in areas of the right occipital lobe, and when attention was directed to the right, there was more activity in areas of the left occipital lobe (Figure 9.25). Note that the point of this study is to demonstrate that directing attention alters neural processing in visual areas of the brain. Thus, attention can affect visual processing. The Neuropsychology of Attention One of the most fascinating, albeit tragic, neuropsychological conditions arises when there is damage to the orienting attention network in the parietal lobe. When a stroke FIGURE 9.24 Attentional Networks and V4 Rhesus macaques fixated on a central marker but attended to stimuli to the left or right of that marker. Moran and Desimone (1985) found that when attention is directed to the effective stimuli, the V4 cell responds strongly. Receptive field of V4 neuron Ineffective stimulus Direction of attention Response of V4 neuron Response of V4 neuron Effective stimulus Fixation point Direction of gaze 276 Sensation and Perception or some other neurological insult affects the right posterior parietal lobe, a condition called hemifield neglect or unilateral visual neglect may arise. This condition almost always occurs when the right parietal lobe is damaged, leading to a deficit in the left visual world. In a handful of patients, one can see neglect of the right side of the world from left parietal damage, but this is much less common than a left-field deficit from right parietal damage. We define hemifield neglect or unilateral visual neglect as a condition in which a person fails to attend to stimuli on one side of the visual world (usually the left) as a consequence of neurological damage to the posterior parietal lobe. What are the consequences of not attending to half of the visual world? It may not seem so strange at first, but it has devastating consequences. For example, male patients may shave only one side of their faces. Patients may neglect the care of the left sides of their bodies. In one case described by Sacks (1985), a patient with neglect rejected the fact that the left side of his body belonged to him. Sensitive tests can show that these patients have not lost the ability to see in their left visual world, but they no longer attend to stimuli in the left world. When pressed to respond to an item in the left visual world, they may identify it, but under normal circumstances, they simply ignore it. It is as if they are no longer interested in the left world or, more properly speaking, that they can no longer direct attention to objects in the left visual world. One of the tests of neglect is to ask patients to copy various drawings. For exam- ple, when asked to reproduce a clock face, a patient may put all the numbers of the clock on the right side of the page (Figure 9.26). When shown a picture of a house, the patient draws only the part of the house on the right side of the paper. When asked to draw her own face, a participant will draw only the side she sees on the right side of the mirror. One striking demonstration that neglect is an attentional problem and not a visual problem comes from a patient in Italy studied by Bisiach and Luzzatti (1978). Their patient was the editor of a major newspaper in an Italian city and a highly educated man. After his initial hospitalization, he was able to resume his duties with the newspa- per. But he continued to visit the lab to be tested by Bisiach and Luzzatti. Asked to imag- ine himself walking down the main square from north to south in his native city in Italy, FIGURE 9.25 Shifts of Attention Within the Occipital Lobe When attention was directed to the left visual world, there was more activity in the right occipital lobe, and when attention was directed to the right visual world, there was more activity in the left occipital lobe. Shift left to right Shift right to left Right hemisphere Left hemisphere 0.30 0.15 −0.15 −0.30 −8 8 12 16−4 4 Time (sec) Left Extrastriate Right Extrastriate % s in g le c h a n g e % s in g le c h a n g e 0 −8 8 12 16−4 4 Time (sec) 0 0 0.30 0.15 −0.15 −0.30 0 Hemifield neglect (unilateral visual neglect): a condition in which a person fails to attend to stimuli on one side of the visual world (usually the left) as a consequence of neurological damage to the posterior parietal lobe 277 Chapter 9: Visual Attention the patient described the monuments and buildings only on the west side (his right). When asked to imagine himself walking up the main square from south to north, the patient described the sites only on the east side (now on his right). Thus, there was nothing wrong with his visual memory. He successfully retrieved the buildings on each side of the main square during the tests. However, when engaged in this mental walk, he attended only to those buildings that would have been on his right. He consistently ignored the left side of the square, regardless of whether “left” meant east or west. Note that he was doing this in the lab, not actually in the square. Thus, he was showing neglect for the left side of his visual images rather than visual perception. Thus, the imagery deficit mirrored the perceptual deficit. Left untreated, the symptoms of uni- lateral visual neglect will persist and can be quite devastating to a patient’s long- term health and well-being. However, patients show improvement with inter- vention and treatment. One of the more successful treatments is to require patients with hemifield neglect to wear prism glasses, which shift the left visual world into the right visual world, where they will attend to it. In some cases, one can, using a series of prisms with different refractive properties, shift the left visual world leftward and see if the patients continue to attend to it (Jacquin-Cortois et al., 2013). Other treatments have been tested as well. Pavlovian conditioning has been used to increase patients’ attention to the left visual world (Domínguez-Borràs, Armony, Maravita, Driver, & Vuilleumier, 2013), and there are studies showing that listening to music can also allow some patients to focus more on the left visual world than if they were not listening to music (Chen, Tsai, Huang, & Lin, 2013). With these treatments, many patients with unilateral neglect will see a diminishment of the symptoms and a chance to resume their normal lives. Actual patients with unilateral visual neglect can be seen in ISLE 9.10. Bálint’s Syndrome Bálint’s syndrome is a rare condition in which function in both the left and the right posterior parietal lobes has been compromised. Patients with this condition have a limited ability to localize objects in space. This results in difficulty grasping for objects, probably its most salient symptom. In addition, patients with Bálint’s syndrome seldom move their eyes. As a consequence of this deficit, they have a condition called simult- agnosia. Simultagnosia is a deficit in perceiving more than one object at a time. Thus, they focus on the one object that is presented directly in front of them and ignore other stimuli. This makes them similar to patients with unilateral neglect, except that patients with Bálint’s syndrome ignore both the left and the right visual world. Actual patients with Bálint’s syndrome can be seen in ISLE 9.11. ISLE 9.10 Hemifield or Unilateral Neglect ISLE 9.11 Bálint’s Syndrome Bálint’s syndrome: a neurological condition caused by damage to both left and right posterior parietal lobes Simultagnosia: a deficit in perceiving more than one object at a time FIGURE 9.26 Hemifield Neglect (Unilateral Visual Neglect) Patients with this neurological condition ignore one half of the visual world (usually the left). These pictures show how patients with hemifield neglect interpret the world. They understand that a clock should have 12 numbers, but because they ignore half of the world, they put all 12 numbers on one side of the clock. 12 11 10 9 8 12 6 111 210 39 48 57 6 1 2 3 4 5 7 278 Sensation and Perception TEST YOUR KNOWLEDGE 1. Find a diagram of the brain and indicate both the orienting and executive networks of attention on it. 2. Classify each of the disorders of attention in terms of their deficits and whether the orienting or executive functions of attention are disrupted. DEVELOPMENTAL ASPECTS OF VISUAL ATTENTION 9.6 Identify what factors lead to the development of visual attention as we grow up. When does the ability to attend develop? Are young infants born with the ability to select locations in space they wish to attend to, or are newborns caught in a world of stimuli without this ability, the “blooming, buzzing confusion” of William James (1890)? Is there an age at which we learn to covertly direct attention to a point in space other than the one we are looking at? There are many pertinent questions in the devel- opment of visual attention. We examine only a small sample of the possible answers to these questions here. Attention in very young infants is determined by how long they can maintain a gaze at an interesting stimulus, whether that stimulus is a colorful collage or an image of their mother (Figure 9.27). As most parents suspect, it turns out that at quite a young age, infants can maintain the direction of gaze for a considerable time, indicative of attentional control (Colombo, 2002). One means of studying attention in young infants is with a technique called the oddball procedure. In this procedure, an infant is shown a series of related objects (e.g., different kinds of balls). However, after a number of balls have been shown to the infant, a novel object from a different category is shown to the infant (e.g., a stuffed animal). In older children, the novel object attracts attention. In infants, such attention would be called stimulus orienting, as the infant focuses on the novel object. This procedure can then be combined with neuro- imaging techniques such as using the event-related poten- tial analysis of EEG recording (Richards, Reynolds, & Courage, 2010). Richards et al. (2010) compared infants who were alert and those who were not (i.e., sleepy). They also compared infants at 4, 5, and 7 months of age in an oddball paradigm. Alert babies and older babies showed greater activity as measured by electroencephalography. In addition, alert babies, regardless of age, showed greater EEG activity when presented with the oddball item. Thus, infants as young as 4 months orient toward novel stimuli (Richards et al., 2010). Other recent neuroimaging studies suggest that ori- entation in the oddball paradigm is accomplished by the same areas of the brain that are associated with selective attention in older children and adults. These brain regions include areas in the extrastriate cortex as well as the pos- terior parietal lobe. Richards et al. (2010) superimposed the EEG activity (discussed in the previous paragraph) FIGURE 9.27 Infants and Attention Do young infants have the ability to direct their attention to specific objects in the environment? © iS to ck ph ot o. co m /n m ax fie ld 279 Chapter 9: Visual Attention on magnetic resonance images of infant brains to con- firm this (Figure 9.28). Figure 9.28 shows the areas of the brain active in an infant brain during the oddball task. Many adults complain that if a television set is play- ing, they have difficulty concentrating on their jobs or on their studies. Indeed, we are sure that you have heard professors advise you not to study with the television on. However, infants may function a bit differently in this regard. Setliff and Courage (2011) examined infants at 6 and 12 months of age during play with novel toys. The infants played with the toys while a television was either playing or not playing in the background. They found lit- tle difference in the attention infants gave to the toys and very little difference in directed glances at the television during the session. Thus, it is likely that background noise provided by a television is less distracting to infants than it is to adults. In this one regard, we might conclude that infants are better at selective attention than their older family members. Certainly, we know from the literature on aging that older adults have a harder time inhibiting responses to irrelevant stimuli than younger adults, and it looks as if the youngest members of our species are even better. TEST YOUR KNOWLEDGE 1. Assess the oddball paradigm and how it can be used to tell us about the development of visual attention. 2. Track the changes in our ability to block out background stimulation from childhood through adulthood. FIGURE 9.28 Neuroimaging Attention in the Infant Brain Activity includes brain regions such as the extrastriate cortex and the posterior parietal lobe. EXPLORATION: Awareness and Visual Consciousness For a philosopher, one of the fundamental questions about perception is the following: How can a physical system composed of brain, eyes, and neurons and made of organic materials, such as proteins and fats, produce con- scious visual experiences, such as the experience of seeing green or the joy in watching a cat swish its tail when it wakes up from a nap (Figure 9.29)? Philosophers of mind argue endlessly whether we can account for awareness and consciousness with purely material explanations, even though they acknowledge that without brains and eyes, no seeing would ever occur. This paradox—that the neural processes of seeing do not resemble in any way the subjective experience of seeing—has led neurosci- ence to examine if we can find the neural correlates of consciousness. Vision and attention provide a great laboratory for looking at the neural correlates of consciousness, because clearly FIGURE 9.29 A Sleepy Cat Many cat owners will smile at the relaxed pose and poise of the cat in this photo. Why does seeing a cat feel the way it does? © iStockphoto.com /chictype 280 Sensation and Perception when you are looking at and attending to an object, such as a cup of chocolate mousse, you are conscious of that object. But as we have seen in this chapter, with phenom- ena such as change blindness, inattentional blindness, and unilateral visual neglect, that you are conscious of seeing the mousse does not mean that you will consciously per- ceive the moose and the mouse, approaching from different directions, which just might attempt to eat your mousse. Thus, in this section, we will examine what two phenom- enal experiences, perceptual bistability and blindsight, can tell us about awareness and vision. Perceptual Bistability Perceptual bistability is one of those phenomena you must see before you can believe, understand, or study. So first go to ISLE 9.12 to see it and experiment with it. You can also use your pair of three-dimensional (3D) glasses to inspect Figure 9.30. The anaglyph here pro- vides an opportunity to see a bistable image. In examining Figure 9.30, note how first you see the cat’s face, and then it fades out and is replaced by the flower, only to have the flower fade out and be replaced by the face again. This is the essence of percep- tual bistability. We can define perceptual bistability as the phenomenon in which a static visual image leads to alternating percep- tions. Perceptual bistability occurs in a number of com- mon illusions, such as the Necker cube, the rabbit–duck image, and the faces–vase image, all of which can be seen in Figure 9.31. Each of these figures creates a paradox for our visual system. For example, in the Necker cube, what we really see are some lines on a two-dimensional surface, namely, the paper of your book or the screen of your computer. However, the lines suggest a 3D cube, and we tend to interpret this figure in three dimensions. However, there are two possible 3D interpretations of this two-dimen- sional figure. So, our perceptual systems may flip back and forth from one to the other. Of interest here is that the exact same physical stimulus creates two separate percep- tual experiences. One of the most striking examples of perceptual bistabil- ity comes from a phenomenon known as binocular rivalry. Binocular rivalry occurs when a separate image is presented to each eye. Figure 9.30 illustrates the perceptual bistability that arises from binocular rivalry. If you look at the image through just the red filter of a pair of 3D glasses, you see only the “red” image, the cat’s face. If you look through only the blue filter of the 3D glasses, you see only the “blue” image, the flower. Thus, when you are using both filters, one over each eye, each eye is getting a separate image. Binocular rivalry illustrates a number of important points about aware- ness in vision. First, our visual system is set up to see a single perception of the world, rather than a separate perception from each eye. Thus, when each eye is seeing a completely FIGURE 9.30 Perceptual Bistability and Binocular Rivalry A bistable image anaglyph. Using a pair of anaglyph glasses, you can look at this image. You should sometimes see one image, sometimes the other. However, if you close one eye and then the other, while using the anaglyph glasses, you can clearly see each image. FIGURE 9.31 Famous Bistable Images Each of these drawings has two different interpretations. Most people report seeing one interpretation and then the other but never both simultaneously. Perceptual bistability: a phenomenon in which a static visual image leads to alternating perceptions Binocular rivalry: a phenomenon that occurs when a separate image is presented to each eye ISLE 9.12 Perceptual Bistability 281 Chapter 9: Visual Attention different image, rather than creating a double image of the world, it systematically inhibits one perception or the other. Second, it also demonstrates the top-down processing aspect of perception. We usually see the more “important” image more than the less important image. Thus, we usually see a face longer than another stimulus in the display, in this case the flower, in Figure 9.30 (Sandberg et al., 2013). In an early study with fMRI, Tong, Nakayama, Vaughan, and Kanwisher (1998) examined two regions of the brain during binocular rivalry. They presented an image that produces binocular rivalry similar to the image depicted in Figure 9.30. They then focused their fMRI scanner on two regions of the brain, the fusiform face area, which we know is involved in face processing, and another nearby area called the parahippocampal area, which responds well to specific objects such as houses. While participants were being mon- itored by the fMRI scanner, they indicated whether they were visually conscious of the face or the house. They were asked to press one button to indicate that they were con- scious of the face and another button to indicate that they were conscious of the house. Amazingly, when the partic- ipants perceived the face, Tong et al. showed that partici- pants had more activity in the fusiform face area, but when participants perceived the house, they had more activity in the parahippocampal region (Figure 9.32). Subsequent studies showed that lower areas in the visual cortex also change responses, depending on which image is being perceived, including V1 and the lateral geniculate nucleus (Meng, Remus, & Tong, 2005). Thus, activity in these areas tracked the participants’ visual consciousness rather than the physical nature of the stimuli (see Tong, Meng, & Blake, 2006). Blindsight Recall blindsight from Chapter 4. Blindsight refers to the residual ability to make visual responses when a patient is subjectively blind in certain regions of his visual field. This means, paradoxi- cally, that patients with this condition are making visual responses to stimuli they cannot see, that is, to which they are not aware. Refer back to Chapter 4 and the story of Patient T.N., a medical doctor who had two strokes less than a month apart in the early 2000s. The strokes caused permanent and complete blind- ness (Buetti et al., 2013). He has no conscious awareness of seeing, yet he can use his vision. You can read more details about his condition in Chapter 4 and see a video of his navi- gating a hallway strewn with obstacles in ISLE 9.13. Various studies have shown he can point in the proper direction of objects on a computer screen over 75% of the time (Buetti et al., 2013) and even distinguish the emotions of faces presented to him (Pegna et al., 2005). So, the question arises, how can someone make visual responses without having the conscious awareness of seeing? At this point, you should be quite familiar with the anatomy of the visual system. You may remember from Chapter 4 that the retinae of the eyes project to many different areas in the brain. The majority of retinal axons project to the lateral geniculate nucleus and from there to the occipital lobe. It is likely that this is the path- way that leads to conscious visual experience. Indeed, it is this pathway that is damaged in T.N. and others who have blindsight. However, the retinae also project to other areas of the brain. Weiskrantz hypothesized that the responses made by patients with blindsight might be the result of behavior produced by these alternate routes (see Weiskrantz, 2009). In particular, Weiskrantz focused on the route from the retinae to the superior colliculus. The superior colliculus is a region of the brain instru- mental in making visually guided head movements as Rivalry Time from reported perceptual switch (s) 1.0 0.8 0.6 0.4 0.2 0.0 −8 −4 0 House Face FFA FFAPPA PPA 4 8 12 % M R s ig n a l 1.0 0.8 0.6 0.4 0.2 0.0 −8 −4 0 Face House 4 8 12 % M R s ig n a l FIGURE 9.32 Bistability in the Brain (Tong et al., 1998) When participants perceived the face, there was more activity in the fusiform face area, but when participants perceived the house, there was more activity in the parahippocampal region. ISLE 9.13 Navigation in Blindsight 282 Sensation and Perception well as eye movements. Weiskrantz hypothesized that because this route is still intact in patients with dam- age to V1, the superior colliculus can guide their visual responses, even though the route that produces con- scious vision, that is, V1, is damaged. Thus, the superior colliculus can allow visual responses in the absence of seeing (Figure 9.33). Interestingly, research now shows that the superior col- liculus projects axons to higher areas of visual process- ing in the occipital lobe. So, the visual responding seen in patients such as D.B. and T.N. may be the result of this connection between the superior colliculus and the occipital cortex, rather than just the superior colliculus alone. However, even if this is the case, this route, though strong enough to allow visual response, is not sufficient to create a sense of conscious visual experience in these patients. TEST YOUR KNOWLEDGE 1. Interpret what perceptual bistability tells us about our general apparent stable perception of the world. 2. Formulate a description for the different functions of those pathways involved in blindsight and the visual pathways that travel through the cortex. Optic chiasm Optic tract Superior colliculus Lateral geniculate nucleus Damaged pathway V1 FIGURE 9.33 The Pathways in Blindsight In people with blindsight, the route from the retinae to the lateral geniculate nucleus to V1 has been damaged. It is this route that supports conscious seeing. However, other routes, such as the link from the retinae to the superior colliculi, may support visual response in the absence of visual consciousness. APPLICATION: Distracted Driving Driving is a task that demands sustained attention. We drive for extended periods of time and the situation is always changing. Some truck drivers, however much it is discouraged, may drive for more than 15 hours without stopping. As we drive, we must scan the road for what other drivers are doing, changes in direction on the road, and even stoplights going from green to yellow to red. The task is complicated enough when driving is all that is required. We can be distracted by other happenings in the car, such as conversations with passengers, our ste- reos playing music, and even making sure we do not run out of gas. But with all of our new devices, especially cell phones, we have a lot of new ways to be distracted while we drive. We simply are more distracted during driving than we used to be. Overall, this distracted driving has grown into a major cause of accidents, particularly those that cause injuries or fatalities. In the years from 2005 to 2009, the overall rate of fatalities from car crashes fell significantly while the rate of fatalities due to distracted driving increased 22% (Coben & Zhu, 2013). This period of time witnessed the introduction of our beloved smart- phones. Beyond crashes, it appears that distracted driv- ing impedes general traffic flow as well (Stavrinos et al., 2013). Using a simulator, drivers who were distracted var- ied speed more, changed lanes significantly fewer times, took longer to complete the scenario, and were passed by more (simulated) vehicles. These behaviors indicate a driver who, even when not making traffic more danger- ous, is still clogging up the roadway and not helping us reach our destination. Several of the concepts that we have discussed in this chapter can help us to understand why distracted driving is such a concern. Overt attention tells us that attention 283 Chapter 9: Visual Attention is usually located on the fovea in the direction of gaze. While looking at the screen of our phone, we are not looking at the road, and features of the road and traffic can easily be missed. Although we can direct attention to other locations than our gaze (covert attention), it does not appear to help us in driving. Moreover, cell phones are highly salient objects, so they draw our atten- tion. There is even some evidence that merely having the phone handy serves to draw our attention and can negatively impact our driving (Thornton, Faires, Robbins, & Rollins, 2014). Cell phones seem to have a strong attentional capture. Even after we finish with the phone, we cannot imme- diately switch back to processing the road. Attentional blink tells us that there is a time lag from processing one stimulus before we can process the next. We learned that it appears that experienced video gamers might have a smaller attentional blink (Kristjánsson, 2013). It seems reasonable that we might find that video gamers might also do better in distracted driving situations. Rupp, McConnell, and Smither (2016) examined this very idea using driving simulators. It is a lot safer to study driving whenever possible using a driving simulator. Researchers can control the traffic and any distractions in simulators without risk to the participant or anyone else. In this experiment, one of the variables the researchers exam- ined was the number of times the drivers did not stay in their lanes, what they called lane deviations. Interestingly, video gamers showed fewer lane deviations during non- distracted driving, when there were fewer irrelevant stim- uli present, than during distracted driving. It seems that video game experience helps us sustain our attention over a prolonged period of time but not deal with distractions (Rupp et al., 2016). One interesting study examined changes in the brain during distracted driving. With visual attention, we cov- ered two types of networks. There is the orienting atten- tion network, centered in the parietal lobe, that allows us to direct our visual attention. When we think about driving, this network will help us study the road to make sure we follow the upcoming curves or examine the mirrors to see what is going on behind us. Then there is the executive action network, which gives top-down control to attention. When applied to driving, the exec- utive action network can help us to respond to driving as opposed to other tasks available. Think about traffic suddenly becoming busy. You have been having a great conversation with your fellow passengers in the car and suddenly you drop out of the conversation. The execu- tive action network would help you put your driving over your conversation. Schweizer and colleagues (2013) asked their participants to drive in a driving simulator while in a device that mea- sured fMRI. They used two basic conditions, a normal driving condition and a distracted driving condition. To simulate the distraction that cell phones cause, the partici- pants needed to answer general knowledge questions in the distracted driving conditions, whereas in the normal driv- ing condition, this was not required. Schweizer et al. then compared the fMRI patterns across the driving conditions. In the normal driving condition, the posterior part of the brain was most active, largely overlapping with the orient- ing attention network. However, during distracted driving, the activity in the brain moved to the forebrain, where the executive action network is located. It seems that during distraction we take activity away from those networks that help us guide our vision where we need it and move it to areas that suppress habitual driving responses and allow us to use our cell phone. There might be some hope in the future. A study by Wang, Chen, and Lin (2014) used electroencephalography to measure brain activity during distracted and normal driv- ing. They were able to identify quite reliably when drivers were distracted or not by looking at the brain activity over the frontal lobe and left motor cortex. The frontal area matches what was observed by Schweizer et al. (2013), providing nice confirmation for both studies. Perhaps in the future there might be a device that can help alert us when we are not properly paying attention to our driving. Or perhaps driverless cars really are the way to go. TEST YOUR KNOWLEDGE 1. Classify the attentional functions, for example, attentional blink and visual capture, that play a role in making distracted driving such a danger. 2. Examine how distracted driving is related to the orienting and executive networks of attention. Sensation and Perception284 CHAPTER SUMMARY 9.1 Explain selective attention and divided attention, which help us understand our scanning of the world. Alertness is a state of vigilance. When alert, we are awake, mindful, and likely scanning our surroundings. Awareness is actively thinking about something, which can be either physically present or just in our imagination. Attention is a set of processes that allow us to select or focus on some stimuli. Attention can also be sustained or temporary. Selective attention can be defined as the pro- cesses of attention that allow us to focus on one source when many are present. Divided attention is the process of attending to multiple sources of information. 9.2 Illustrate how we move attention around a scene both by using eye movements and by using our peripheral vision. Generally, we align our attention with the direction of our gaze. A question we can ask, however, is whether it is pos- sible to attend to spatial locations other than the location we are looking at. Posner and his colleagues investigated this phenomenon. A cue is given that a target will occur in a particular region of space. In most trials, the target appears in the cued location, but in some it does not. Even though participants cannot look at the cued area, their response times are faster when the cues are accurate. This tells us that we can devote attention covertly, that is, direct atten- tion to a location in space we are not looking at directly with our foveae. This helps us understand the distinction between overt and covert attention. Overt attention is when our atten- tion is directed with our gaze, and covert attention is when our attention is directed toward our periphery. Inattentional blindness is a phenomenon in which people fail to perceive an object or event that is visible but not attended to. Chabris and Simons (1999) conducted the famous gorilla experiment to demonstrate that salient stimuli can actually be missed when the focus of attention is elsewhere. 9.3 Assess the nature of attentional capture and how it works in our visual system to have stimuli capture attention. Stimulus salience means that some objects in the environ- ment attract our attention. Attentional capture occurs when a salient stimulus causes us to shift attention to that stimu- lus. Both a history of previous reward and relevant semantic meaning tend to capture our attention. Visual search means looking for a specific target in an image with distracting objects. Feature search means searching for a target in which the target is specified by a single feature. Conjunction search means searching for a target in which the target is specified by a combination of features. Feature integration theory stipulates that some features can be processed in parallel and quickly prior to using attentional resources. Other visual characteristics require us to use attention and are done serially and therefore less quickly. 9.4 Examine attentional blink and change blindness as ways of helping us understand how attention changes over time. Rapid serial visual presentation (RSVP) is an experimental paradigm in which a series of stimuli appear rapidly in time at the same point in visual space. Participants must direct their attention across the time domain rather than the space domain. Findings show that people can respond quickly to a particular stimulus when it occurs in time. Attentional blink is the tendency to respond more slowly or not at all to the sec- ond target in an RSVP task when the second stimulus occurs within 500 ms of the first stimulus. Repetition blindness is the failure to detect the second target in an RSVP task when the second target is identical to the first one. Like attentional blink, it occurs when the second target is presented 500 ms or less after the first target. Change blindness is the difficulty we experience in detecting differences between two visual stim- uli that are identical except for one or more changes to the image. In a surprising finding, Simons and Levin (1997) found that participants could even be change blind to the substitu- tion of one person for another. 9.5 Explain the differences between the orienting attention network and the executive attention net- work and their roles in directing visual attention. The orienting attention network (or dorsal attention network) allows us to engage in visual search and direct our visual attention to different locations in visual space. The orienting attention network is located primarily in the parietal lobe. The executive attention network focuses on attention as the inhibition of habitual responses and the top-down con- trol of attention. The executive attention network is found in the frontal lobe. Moran and Desimone (1985) found that at a physiological level, attention can affect the processing of visual information. Attention acts to decrease the activity of a cell when the effective stimulus for that cell is not being attended to. Hemifield neglect, or unilateral visual neglect, is a condition in which a person fails to attend to stimuli on one side of the visual world (usually the left) as a con- sequence of neurological damage to the posterior parietal lobe. Bálint’s syndrome is a neurological condition caused by damage to both the left and the right posterior parietal Chapter 9: Visual Attention 285 lobes. Simultagnosia is a deficit in perceiving more than one object at a time. Perceptual bistability is a phenomenon in which a static visual image leads to alternating perceptions. Binocular rivalry occurs when a separate image is pre- sented to each eye. Binocular rivalry results in perceptual bistability as viewers see the image presented to one eye for a while before the perception shifts to what the other eye is viewing. Blindsight is the residual ability to make visual responses when a patient is subjectively blind in certain regions of his or her visual field. Patients will say that they are blind in regions of their visual fields but make responses anyway. It is likely that the blindness is caused by damage to area V1 of the occipital cortex but that the residual vision is a function of intact structures, such as the superior colliculus. 9.6 Identify what factors lead to the development of visual attention as we grow up. Infants show from quite a young age the ability to maintain the direction of gaze, indicating some attentional control. One way attention is studied in infants is with the oddball procedure. If infants are shown a series of objects from the same cate- gory and then a novel object is presented, the oddball attracts attention. This procedure has shown that infants as young as 4 months will orient toward the novel object (Richards et al., 2010). These same researchers also found that some of the same regions used to direct attention in older children are used by infants. It has also been found that, compared with older children and adults, infants do not seem as distracted by external stimuli, say a television (Setliff & Courage, 2011). REVIEW QUESTIONS 1. What is the difference between selective attention and divided attention? What function does attention serve in perception? 2. How did Posner and his colleagues study covert attention? Describe their experiment and its results. What do these results tell us about covert attention? 3. What is attentional capture? Describe an experiment reviewed in the textbook that examined features of attentional capture. 4. What is the difference between change blindness and inattentional blindness? Describe an example of each. Why are change blindness and inattentional blindness considered errors of attention? 5. What is feature integration theory? How does it explain feature searches and conjunction searches? 6. What is the rapid serial visual presentation (RSVP) paradigm? How is it conducted in the lab, and what aspect of attention does it measure? 7. What are the orienting attention network and the exec- utive attention network? What area of the brain is each located in? What aspects of attention do they serve? 8. What is unilateral neglect? What are its symptoms? What kind of brain damage causes it? 9. What is binocular rivalry? How is it measured, and what principles of awareness does it demonstrate? 10. What is blindsight? Describe the deficits and pre- served function in the patient T.N. What is typically thought of as the neurological basis for the preserved function in blindsight? 11. How does our understanding of visual attention help us explain the safety issues associated with cell phone use during driving? PONDER FURTHER 1. You are watching your favorite team playing your favorite sport. The official (referee, umpire, etc.) makes a call that you disagree with, vigorously. We have all been there. Take the material from this chapter and evaluate the attention of both yourself and the official to consider why you both see the situation so differently. Without casting blame, consider why you might see these situations so differently. Consider your location, stimuli, and internal factors, and even your orienting and executive networks. 2. You are an advertiser or marketer. How can you apply the material from this chapter to enhance the chance of visual capture by your advertisement, announce- ment, or solicitation? Now imagine you are a person receiving this information. How does what has been covered help you to internally choose your stimuli for attention? Answer for Figure 9.19 : In one image, there are three yellow and gold flowers in the lower left-hand corner. In the other image, there are only two yellow and gold flowers in the lower left-hand corner. Sensation and Perception286 KEY TERMS Attention, 257 Attentional blink, 272 Attentional capture, 265 Automaticity, 258 Bálint’s syndrome, 277 Binocular rivalry, 280 Change blindness, 269 Conjunction search, 267 Covert attention, 259 Divided attention, 258 Executive attention network, 273 Feature integration theory, 267 Feature search, 267 Hemifield neglect (unilateral visual neglect), 276 Inattentional blindness, 262 Orienting attention network (dorsal attention network), 273 Overt attention, 260 Perceptual bistability, 280 Rapid serial visual presentation (RSVP) paradigm, 271 Repetition blindness, 272 Selective attention, 258 Simultagnosia, 277 Stimulus onset asynchrony, 260 Stimulus salience, 265 Visual search, 266 Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review, study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning. Learning Objectives Digital Resources 9.1 Explain selective attention and divided attention, which help us understand our scanning of the world. The Neural Basis of Selective Attention: Cortical Sources and Targets of Attentional Modulation Identifying and Remediating Failures of Selective Attention in Older Drivers The Science Behind Sleight of Hand 9.2 Illustrate how we move attention around a scene both by using eye movements and by using our peripheral vision. How We Pay Attention The Invisible Gorilla (Featuring Daniel Simons) Seeing the World as It Isn’t—Daniel Simons 9.3 Assess the nature of attentional capture and how it works in our visual system to have stimuli capture attention. Intersensory Redundancy Guides the Development of Selective Attention, Perception, and Cognition in Infancy We See More Than We Can Report: “Cost Free” Color Phenomenality Outside Focal Attention What We Really See When We Go See a Movie 9.4 Examine attentional blink and change blindness as ways of helping us understand how attention changes over time. The Attentional Blink Reveals the Probabilistic Nature of Discrete Conscious Perception Change Blindness: Theory and Consequences Optical Illusion: Car Wheels Going ‘Round 9.5 Explain the differences between the orienting attention network and the executive attention network and their roles in directing visual attention. The Blind Woman Who Sees Rain, but Not Her Daughter’s Smile A Visual Neglect Patient Excerpt From Through the Wormhole With Morgan Freeman: Interview With Beatrice de Gelder 9.6 Identify what factors lead to the development of visual attention as we grow up. Link Seen Between Babies’ Sight, Language Development How Do You Ask a Preverbal Infant What She Can See? Victor Habbick Visions/Science Source 10The Auditory System DNA Illustrations/Science Source LEARNING OBJECTIVES 10.1 Discuss the physical nature of sound and how it correlates with the perceptual nature of sound. 10.2 Describe the basic anatomy of the ear and the difference between the outer, middle, and inner ear. 10.3 Examine the nature of hearing loss and what can be done to improve hearing in hearing-impaired individuals. INTRODUCTION “Friends, Romans, countrymen, lend me your ears.” So wrote William Shakespeare in his play Julius Caesar (1599). Of course, Caesar meant for his audience to listen to him, not actually allow him to borrow their ears. There was some confusion about this in the mind of one of your authors when he was a young boy and first heard this quotation from the play. But as is often the case with Shakespeare, he summarized quite a bit in one clever line. For all but a small population of sign language speakers, sound means communication and the ability to talk and listen to others. To lend someone your ears means to listen to the language sounds that he or she is making. Indeed, one of the crit- ical functions of hearing is to perceive and interpret the sound of the human voice. It is hard to imagine a world in which we could not hear one another talk, and for billions of people around the globe, sound and hearing also mean music, which brings us great enjoyment and peace of mind. Indeed, people who have gone deaf after a lifetime of hearing most often miss the ability to hear music. Think of all the sounds you hear at any given moment. You may hear the whirring of the washing machine, the whining of the air conditioning, the distant buzz of cars on the road, a lawn mower from down the street, and the gentle lapping sounds of your cat drinking from its water bowl. Our ears are always open, and thus we hear these sounds all the time. This contrasts with vision, in that you can close your eyes. In some cases, sounds such as the ones just described may become distracting. In this case, you may create your own world of sound by putting on headphones or ear buds. In this case, listening to your favorite music allows you to concentrate on the sounds you want to hear, not simply those present in the environment. Alternatively, you may put in ear- plugs to drown out external sound. Although this may succeed in dampening unwanted sound, many people find the silence of this to be strange (Figure 10.1). Moreover, sound comes at us from all directions. We see only in front of us, but we hear from all 360 ISLE EXERCISES 10.1 The Sound Stimulus 10.2 The Speed of Sound and the Sonic Boom 10.3 The Decibel Scale 10.4 Frequency and Pitch on a Piano 10.5 Frequency Response of the Ear 10.6 Timbre and Musical Instruments 10.7 Fourier Analysis in Audition 10.8 Missing Fundamental 10.9 Timbre and Overtones 10.10 Phase and Cancellation 10.11 The Middle Ear 10.12 The Basilar Membrane and Sound Stimuli 10.13 The Traveling Wave 10.14 Place Code Theory 10.15 The Basilar Membrane and Fourier Analysis 10.16 Transduction and Hair Cells 10.17 Temporal Code Theory 290 Sensation and Perception degrees. This is likely why wailing sounds, such as a siren, and not flashing lights indicate emergencies that are out of visual range. In sum, we live in a world engulfed in sound. As with every perceptual system, the essence of hearing is transforming information available in the environment, in this case, sound pressure waves, into a perception we can use to understand the world and guide our actions. As we work through the sensory processes involved in hearing and the per- ceptual processes involved in speech perception and music per- ception, you can note both the similarities and the differences between audition and vision. They share similar physiological mechanisms, but these mechanisms must be modified for the specific modality they are responsible for. SOUND AS STIMULUS 10.1 Discuss the physical nature of sound and how it correlates with the perceptual nature of sound. In the movie Alien (1979), one of the scariest movies ever made, the tag line is “In space, no one can hear you scream” (Figure 10.2). As effective as this is in building tension before you even go into the movie theater (and maybe eliciting screams in the theater), the statement is true, in the sense that there needs to be a medium, such as air or water, to conduct sound. In the absence of such a medium, as in the near vacuum of space, sound cannot exist. So, before we can discuss the physiological and psychological processes that allow us to hear, we need to understand a little bit about what sound is. The sound stimulus is the periodic variations in air pressure traveling out from the source of the variations. That is, these periodic variations in pressure are the sound wave, and the source of the variations is the object making the sound. Over time, air pressure will increase and decrease slightly, and these small changes in air pressure con- stitute sound to our ears if they occur strongly enough and quickly enough. Sound can also be transmitted through other media, such as water, bones, and train tracks—any- thing that can vibrate. Water transmits sound faster than air does, making it more dif- ficult for a scuba diver to localize a sound underwater than for a person on land. Even someone just swimming in a pool may notice how difficult it is to tell where a sound is coming from. Mostly, humans hear in an air medium, and occasionally in a water medium, but really anything that can reverberate or change in pressure over time can transmit sound. Think of the Old West sheriff listening for the bad guys by putting his ear to the train track. The iron of the rail transmits sound, which tells the sheriff when the train is coming and therefore when the train robbers will be coming. Think of what happens when you clap your hands. The pressure of your hands against each other compresses the air between them, creating the pressure wave. This compression of the air between your hands causes the air molecules to collide with other air molecules in each and every direction around your hands, which then collide with air molecules farther away from your hands, and so on and so forth. As the air molecules are compressed in each area of space, just behind them is a small area in which the air pressure is lower because some of the molecules have been pushed forward. Thus, sound consists of pockets of higher pressure air followed by pockets of lower pressure air. Changes in air pressure propagate outward from the original source of the disturbance, © iS to ck ph ot o. co m /L es lie La ur en FIGURE 10.1 Ear Protection People who do jobs with a lot of noise nearby are often advised to wear ear protection. Ear plugs like this can block out loud noise. Sound stimulus: the periodic variations in air pressure traveling out from the source of the variations M ov ie st or e co lle ct io n Lt d/ A la m y St oc k Ph ot o FIGURE 10.2 Alien (1979) 291 Chapter 10: The Auditory System in this case your hands clapping (Figure 10.3). For a demonstration of this, go to ISLE 10.1a. When you clap your hands, you disturb the air around your hands and initiate the pattern of high- and low-pressure air movements that move out in all directions from the source. This pattern is called a sound wave. Sound waves are the waves of pressure changes that occur in the air as a function of the vibration of a source. Sound waves are like any kind of wave—the wave moves through its medium even though particles within the wave may stay in the same place, just as a standing wave stays put on a river. In the case of sound, the source of the sound is your hands when you clap. The source of the sound could be air blown through a brass instrument or the vibrating of strings on a guitar. Air molecules do not travel very far. They basically just move back and forth from high-pressure to low-pressure areas, but the wave of pressure change moves across space, causing compression (high pressure) and rarefaction (low pressure) across space, sometimes at great distances, depending on how forceful the original sound was. In this wave, there will be a peak high in air pressure and a peak low in air pressure. Although the physics of light are much more compli- cated, sound can be measured in an analogous manner, namely, by mea- suring wavelength and its inverse, frequency. The time between two consecutive high peaks is the cycle of a sound wave. Cycles can be measured in their number per second, also known as frequency (Figure 10.4). Finally, the energy in any sound wave will weaken across time and space. Your clap may sound very loud to you because you are so close to it. But to your friend in another room, it will be audible but not loud. By the time the sound wave gets through the door and out into the street, the energy may have dropped off so much that a person walking by on the sidewalk may not hear it at all. You can see a demonstration of this on ISLE 10.1b. Under normal conditions, at sea level, sound travels very fast, but much more slowly than light. Sound travels at about 344 m/s (761.2 mph). This is a bit faster than a civilian jet plane (a typical jet on its way from Chicago to New York travels at a peak speed of 550 mph). However, we have planes, mostly used by the military, that can equal and exceed the speed of sound. Indeed, the military has planes that exceed the speed of sound by as much as 3 times (“Mach 3” in military terms means 3 times the speed of sound). Such high speeds are avoided for typical civilian aircraft because they leave a tremendous collision of sound, called a sonic boom, behind them when the plane reaches and exceeds the speed of sound (Figure 10.5). Interestingly, because the planes are going faster than the sounds they make, pilots and passengers on such supersonic planes experience an incredibly quiet ride. And, by the way, the Mach of Mach 3 is the same scientist as the person who first observed Mach bands (Chapter 3). ISLE 10.1 The Sound Stimulus FIGURE 10.3 The Propagation of Sound Through a Medium When a person claps, air pressure changes are created. The clap creates a compression of air between the hands, which pushes against the air adjacent to it in all directions. The compression creates a rarefaction of air next to it, which alternates with the compression and propagates through the air. Pressure in small volumes of air molecules CompressionCompression Rarefaction Sound waves: the waves of pressure changes that occur in the air as a function of the vibration of a source Cycle: in a sound wave, the amount of time between one peak of high pressure and the next FIGURE 10.4 Frequency or Wavelength of Sound The vibrations coming from an object, such as hands clapping, can be measured by looking at the pattern of increasing and decreasing air pressure. One cycle is the amount of time it takes for air to return to the same state of air pressure as at an earlier point in the wave or, more technically, the time between two consecutive high peaks. A ir pr es su re Time 1 2 3 4 50 One cycle 292 Sensation and Perception Although the speed of sound is very high, there are times when we may notice a lag between sound and sight. If the source of a sound is 100 meters (m) away, there will be a lag of about 0.3 seconds (s) between when we see the hands clapping and when we hear the sound. Thus, a person watching a sprint race from just behind the finish line will see the gun go off and then hear it a split second later. When we shout across a canyon toward a cliff, the echo is not simultaneous with our shout; it takes time for our shout to reach the distant cliff and for the echo to return. The lag of sound relative to light also accounts for the delay in hearing thunder after we see lightning. Thunder and lightning result from the same event: It is simply that the light reaches your eyes much more quickly than the sound reaches your ears. Interestingly, in water, sound travels about 4 times faster than it does through air (1,483 m/s) (ISLE 10.2). The Relation of Physical and Perceptual Attributes of Sound If you recall, we talked about light as a wave. In examining waves, there are two important measures, their amplitude and their wavelength. Amplitude, for light, means the intensity or brightness of the light, and wavelength dictates the color. Sound has similar attributes—amplitude and frequency (the inverse of wavelength). We also consider the concept of waveform, that is, how different frequencies interact with one other to create complex sounds, which also affects our auditory perception. Each of these physical attributes maps onto a perceptual attribute. Amplitude maps onto loudness, frequency maps onto pitch, and waveform maps onto timbre. To start off our dis- cussion, however, we consider pure tones, that is, sound waves in which air pressure changes follow the basic sine wave for- mat. A pure tone is heard at a particular pitch but does not have the complexity you would expect when hearing a musical instrument (or a voice) play (or sing) that particular pitch. So we consider each of these attributes now (Figure 10.6). To listen to a sample of pure tones, go to ISLE 10.1c. Amplitude and Loudness The amplitude of a sound is expressed as the difference between its maximum and min- imum sound pressures. Thus, as with ocean waves, as any surfer will tell you, taller waves are stronger waves. A strong clap produces a bigger change in air pressure than a mild, not-quite-excited clap. Dropping a hammer results in a ISLE 10.2 The Speed of Sound and the Sonic Boom FIGURE 10.6 Amplitude, Frequency, and Waveform (a) Difference between a low- and a high-frequency sound. The perceptual equivalent of frequency is pitch. (b) Difference between a low- and a high-amplitude sound. The perceptual equivalent of amplitude is loudness. (c) Difference between a simple waveform or pure tone and a complex sound. Complex waveforms contribute to our perception of timbre. Low frequency (low-pitched tone) High frequency (high-pitched tone) Low amplitude (soft tone) Time A ir p re ss u re A ir p re ss u re Time Simple waveform (pure tone) High amplitude (loud tone) Complex waveform (complex sound) A ir p re ss u re Time Normal air pressure A ir p re ss u re Time A ir p re ss u re Time A ir p re ss u re Time A ir p re ss u re Time A ir p re ss u re Time (a) (b) (c) Frequency (perceived as pitch) Amplitude (perceived as loudness) Waveform (perceived as timbre) Pure tone: a sound wave in which changes in air pressure follow a sine wave pattern FIGURE 10.5 This U.S. Air Force F-22 Raptor is capable of exceeding the speed of sound. Its top speed is Mach 1.8 or nearly twice the speed of sound. RE U TE RS /H o N ew 293 Chapter 10: The Auditory System greater amplitude sound wave than does drop- ping a feather. Think about the amount of air that gets displaced when an airplane rushes by. This force of the airplane creates large- amplitude sound waves, which can be heard for miles in all directions. Amplitude has a very clear psychophysical correlate. Loudness is the perceptual experience of amplitude or the intensity of a sound stimulus. High-amplitude (high-intensity) sounds will be heard as loud, and low-amplitude (low-intensity) sounds will be heard as soft. Thus, a person screaming pro- duces greater amplitude of the human voice than does a person whispering. Amplitude is usually measured in decibels. A decibel (dB) is 1/10 of a bel, which is a unit of sound intensity named in honor of Alexander Graham Bell. The decibel scale is a logarithmic scale. This means that the intensity (amplitude) of a sound increases more quickly than the num- bers along the decibel scale. With every 6 dB, the sound pressure doubles. A decibel level of 120 is nearly 1 million times greater in terms of sound pressure than a decibel level of 10. The formula to determine decibel level is given by (this is one of the few formulas we actually include in this book) dB = 10log(p2/p r 2), in which dB is the sound pressure level, p is the measured sound pressure (in another unit called micropascals), and p r is an agreed-upon reference sound pressure. Sound measures of decibels are measures of relative intensity, so it is important to know what both sounds are. In the most usual case, p r derives from the audi- tory threshold that is being used as the com- parison sound. In this case, the measure should be given as dB SPL, where SPL stands for “sound pressure level.” Figure 10.7 gives the approximate decibel SPL levels of a number of common sounds. Sustained exposure to sounds over 85 dB is potentially damaging, and even short 1-second exposures to sounds over 120 dB can result in immediate hearing damage. What this means in terms of our loudness perception is that we can hear sounds across a wide range of amplitudes, from really quiet sounds (the proverbial pin dropping) to sounds so loud that they can damage our ears. In addition, we are remarkably good at detecting differences in loudness. Under quiet conditions at average frequencies (e.g., 2,000 Hz), we can discriminate a difference of just 1 dB. For example, if a series of simple pure tones are played one after the other, at the same frequency, we can detect the one tone that is 1 dB stronger than the other tones. We can perform similarly with complex tones as well (ISLE 10.3). FIGURE 10.7 Amplitude (Loudness) of Common Sounds Decibels (dB SPL) Lawn mower at close range; stereo level of many listeners. Flushing toilet Normal conversation Quiet of�ce level; normal refrigerator sound level Usually suf�cient sound to awaken a sleeping person Empty theater; whispered conversation Sound threshold of a normal young adult Micropascals 100,000,000 130 120 110 100 90 80 70 60 50 40 30 20 10 0 Immediate and permanent hearing loss Jet airplane at takeoff; deafening if no protection Prolonged exposure above 85 decibels can cause noise-induced hearing loss Sound induces pain; even short-term exposure can cause permanent damage. Very loud; sustained sound at this level can cause damage. 10,000,000 1,000,000 100,000 10,000 1,000 100 20 140 150 160Handgun �ring 170 180 Amplitude: the difference between maximum and minimum sound pressures Loudness: the perceptual experience of the amplitude or intensity of a sound stimulus Decibel (dB): a physical unit that measures sound amplitude 294 Sensation and Perception Loud sounds can be dangerous. Prolonged contact with sounds above 85 dB can cause hearing loss in the long run. That is, people who work around such loud sounds or play loud music frequently run the risk of damaging their hearing. This might include the 90-dB sounds of airplanes taking off from the perspective of a baggage handler. This is why people who work on the tarmac are supposed to wear ear protection, to reduce that 90-dB level to something less harmful. Firearms instructors and other people who spend extended time at firing ranges may also experience hearing loss in the long run, even if they wear ear protection. Even with a 30-dB reduction from ear protection, many firearms will still produce sounds that can damage the shooter’s ears. Finally, many people listen to car stereos at sound levels that approach or exceed 100 dB, which can lead to damage in the long run. Even a hair dryer at close range can exceed 85 dB, so be careful. There is a moral here: Be mindful of your daily exposure to such loud sounds. Sounds louder than 120 dB are decidedly painful, and sounds louder than 130 dB will generally result in immediate and per- manent hearing loss. A jet airplane at close range may be 140 dB, hence the need for tarmac workers to protect their ears at all times. Similarly, the firing of many guns is usually in excess of 160 dB at close range, and hence the need for serious hearing protection. Even small-caliber bullets can fire at nearly 140 dB. Thus, firing a gun without ear protection can cause immediate damage to one’s ears, which is a common symptom in veterans of wars. Most stan- dard ear protection for shooters and tarmac workers results in only a net reduction of 30 dB, which puts the shooters and workers out of the range of pain and immediate damage but leaves them high in the zone of sounds that cause long-term damage. TEST YOUR KNOWLEDGE 1. What are the physical dimensions of amplitude and frequency, and what perceptual dimensions do they map onto? 2. Why are loud sounds potentially dangerous to your ears? Frequency and Pitch The frequency of a sound stimulus refers to the number of cycles in the sound wave that occur in 1 second. The perceptual correlate of frequency is pitch. Tones that have low frequencies are heard as being low in pitch, whereas tones that have high frequencies are heard as high in pitch (ISLE 10.4). We define frequency as the number of cycles in a sound stimulus that occur in 1 second. Pitch is the subjective experience of sound that is most closely associated with the frequency of a sound stimulus. Remember that frequency and wavelength are inverses of each other. Frequency is the number of cycles per second, and wavelength is the time course of one cycle. Thus, as the frequency gets larger, the wavelength gets smaller, and vice versa. For arbitrary reasons, we use a wavelength measure when discussing vision but a frequency measure when discussing sound. This is not done just in this book to confuse you; rather, it is the convention in all fields that concern themselves with physical measurements of light and sound. Frequency is measured in hertz (Hz), which is a unit of measure indicating the number of cycles per second, named after the German physicist Heinrich Hertz. A sound that has 200 cycles per second (i.e., 200 ups and downs per second) is said to be 200 Hz. A sound that has 20,000 cycles per second (i.e., 20,000 ups and downs per second) is said to be 20,000 Hz. ISLE 10.3 The Decibel Scale © iS to ck ph ot o. co m /m ar cd uf FIGURE 10.8 It’s a noisy world out there; ear protection can help save you from hearing loss. ISLE 10.4 Frequency and Pitch on a Piano Frequency (sound stimulus): the number of cycles that occur in a second Pitch: the subjective experience of sound that is most closely associated with the frequency of a sound stimulus; related to the experience of whether the sound is high or low, such as the two ends of the keyboard of a piano Hertz (Hz): a unit of measure indicating the number of cycles per second 295 Chapter 10: The Auditory System The psychological equivalent of frequency is pitch. Pitch is the subjective experience of sound that is most closely associated with the frequency of a sound stimulus. Lower frequencies are heard as lower in pitch, and higher frequencies are heard as higher in pitch. Think of the typical man’s voice and woman’s voice. Typically, women have voices with higher pitches than those of men. You can also think of a piano key- board. If you play the note farthest to the left, you are playing a note with a low frequency (27 Hz). Play the note farthest to the right, and you are playing a note with a high frequency (4,186 Hz). Middle C is 261 Hz (Figure 10.9). You can hear how frequency relates to pitch in ISLE 10.4. Children and young adults can hear a range from about 20 to 20,000 Hz (Yost, 2007). As we get into our late 20s and beyond, we lose much of our hearing in the high- est range. By the time one is 40 years old, it is unlikely that frequencies above 14,000 Hz are heard. By the time one is 50 years old, this upper limit may be down to 12,000 Hz. Dr. Krantz describes an experience in his lab when he was a graduate student in which the students complained to the professor about loud, high-pitched, unpleasant sounds being produced by equipment in the lab. The professor, old enough to be deaf in this range, did not hear the sounds, but they were loud and painful to the younger people in the lab. The loss of high-frequency hearing is inevitable, but it may be exacer- bated by exposure to loud sounds when young. The lowest frequencies, however, tend to remain stable with age. To determine your range of frequencies, go to ISLE 10.5. Above 20,000 Hz and below 20 Hz, humans are simply deaf no matter how loud the sound is. Thus, we could play a sound at 30,000 Hz, and we would not hear it, regardless of the decibel level. However, your dog would hear the 30,000-Hz sound. If it was loud enough, the dog just might complain about it, too. In fact, dogs can hear up to about 50,000 Hz. This is the basis of the dog whistle, which makes a sound inaudible to us but perfectly audible to your dog. This allows you to communicate to your dog without other people being aware. Other animals can hear even higher frequencies than that. Bats and dolphins, for example, can hear frequencies up to about 200,000 Hz. These high frequencies are essential for their biosonar systems, which we discuss later in the book. Other animals, such as elephants and humpback whales, can hear much lower frequency sounds than we can. Elephants, for example, use calls as low as 1 Hz for long-distance communication. These low-frequency sounds can travel farther than higher frequency sounds, and whales can make very loud low-frequency sounds that can be heard by other whales many miles away. Losing these high frequencies has few consequences for understanding speech, as even the highest soprano’s frequency is about 1,200 Hz. Similarly, music seldom reaches frequencies higher than 4,186 Hz (the highest note on a piano), which is higher than instruments noted for their high-pitch notes, such as violins and piccolos, can go. Losing the high frequencies interferes with the perception of timbre, our next topic. Waveform and Timbre Imagine the difference between the sounds of a clarinet and a trumpet. A clarinet and trumpet may be playing the same note, such as a B-flat in the same octave (i.e., at the 36 9. 99 41 5. 30 34 .6 48 38 .8 91 46 .2 69 51 .9 13 58 .2 70 69 .2 69 77 .7 82 92 .4 99 10 3. 83 11 6. 54 13 8. 59 15 5. 56 18 5. 00 20 7. 65 23 3. 08 27 7. 18 31 1. 13 46 6. 16 55 4. 37 62 2. 25 73 9. 99 83 0. 61 93 2. 33 11 08 .7 12 44 .5 14 80 .1 16 61 .2 18 64 .7 22 17 .5 24 89 .0 29 60 .0 33 22 .4 37 29 .3 29 .1 35 27 .5 00 Frequency (Hz) N ot e 30 .8 63 32 .7 03 36 .7 08 41 .2 03 43 .6 54 48 .9 99 55 .0 00 61 .7 35 65 .4 06 73 .4 16 82 .4 07 87 .3 07 97 .9 99 11 0. 00 12 3. 47 13 0. 81 14 6. 83 16 4. 81 17 4. 61 19 6. 00 22 0. 00 24 6. 94 26 1. 63 29 3. 67 32 9. 63 34 9. 23 39 2. 00 44 0. 00 49 3. 88 52 3. 25 58 7. 33 65 9. 26 69 8. 46 78 3. 99 88 0. 00 98 7. 77 10 46 .5 11 74 .7 13 18 .5 13 96 .9 15 68 .0 17 60 .0 19 75 .5 20 93 .0 23 49 .3 26 37 .0 27 93 .0 31 36 .0 35 20 .0 39 51 .1 41 86 .0 FIGURE 10.9 Frequency and Pitch A piano keyboard nicely illustrates the relation between frequency and pitch. As one moves to the right from the keys farthest to the left on a piano, frequency increases. The lowest key on a piano is 27 Hz, whereas the highest is 4,186 Hz. As one moves from left to right on the piano keyboard, the notes increase from low in pitch to high in pitch. ISLE 10.5 Frequency Response of the Ear 296 Sensation and Perception same frequency), and yet they still sound different. A trumpet’s sound is pointed and comes right after you, even when played softly. A clarinet seems to have a more subtle sound, which seems to sneak up on you. Composers will call on trumpets often for happy or martial music, whereas the clarinet is often used to express a more sad sound. A piano playing the same note might sound more neutral than either the trumpet or the clarinet. How is it that different instruments playing the same note can sound so different? (To hear the sounds of different musical instruments, you can go to ISLE 10.6.) The difference in sound quality when different instruments play the same note has to do with the concept of harmonics. Harmonics are higher frequencies present in a complex sound that are inte- ger multiples of the fundamental frequency (main frequency). There are a lot of terms in that definition that need explaining. So we will now take a step back and explain the concepts of the fundamental frequency, harmonics, and complex sound. Pure tones are simple sine waves at single frequencies. However, in nature, pure tones are virtually nonexistent. Almost all sounds are complex sounds, which consist of mixes of frequencies. These frequencies combine to form a complex waveform (Figure 10.10). A complex waveform can be broken down into its composite frequencies through a mathematical formula known as Fourier analysis. Fourier analysis is a mathematical procedure for taking any complex waveform and determining the sim- pler waveforms that make up that complex pattern. The simpler waves used are sine waves. When we do a Fourier analysis, we break down a complex sound into its fundamental frequency and its harmonics. The fundamental frequency is the lowest frequency present in the complex sound and the one that determines the perceived pitch of that sound. The harmonics are all frequencies present in the stimulus that are higher in frequency than the fundamental frequency. The fundamental frequency determines the pitch of the sound, but the harmonics provide the timbre that makes the sound of the clarinet different from that of the trumpet or the piano. See ISLE 10.7 for a Fourier analysis of real sound in real time. So now we continue to unpack this new vocabulary. The fundamental frequency determines the pitch. Thus, if the lowest frequency in a complex sound is 440 Hz, we will hear this note as a pitch at concert-tuning A. However, there may also be frequen- cies present at 880 Hz, 1,320 Hz, 1,760 Hz, 2,200 Hz, and so on. These additional frequencies represent the harmonics (Figure 10.11). These harmonics determine the characteristics that give each instrument (or voice, or any sound) a distinct timbre when playing the same note. Observe that the fundamental frequency, though usually louder than its harmonics, need not be the loudest frequency. It is always the lowest frequency present in a sound. Even if the first harmonic is louder than the fundamen- tal frequency, as it is in a clarinet, we still hear the pitch at the fundamental frequency. Indeed, we can remove the fundamental frequency altogether, and we will still hear the pitch at the missing fundamental. We hear the missing fundamental’s pitch because the sequence of harmonics implies its presence (Schwartz & Purves, 2004). ISLE 10.8 illustrates pitch both with and without the fundamental frequency. We hear these differences in the relative strength of harmonics as differences in timbre. Timbre is the musical term that refers to the perceived sound differences between sounds with the same pitch but possessing different higher harmonics. Timbre provides the richness in sound we perceive when we hear a good violinist playing on Harmonics: higher frequencies present in a complex sound that are integer multiples of the fundamental frequency (main frequency) Complex sound: a sound consisting of a mix of frequencies Fourier analysis: a mathematical procedure for taking any complex waveform and determining the simpler waveforms that make up that complex pattern; the simpler waves used are sine waves Fundamental frequency: the lowest frequency in a complex sound, which determines the perceived pitch of that sound Timbre: the perceived sound differences between sounds with the same pitch but possessing different higher harmonics ISLE 10.6 Timbre and Musical Instruments FIGURE 10.10 Complex Waveforms When three notes (green, light blue, and orange) are combined, the result is a complex waveform (dark blue). Fourier analysis can be used to determine how the component waves are combined to create the complex waveform. 60 40 20 −20 −40 −60 0 0 2.5 7.5 10.0 Time (msec) Complex waveform A m p li tu d e ( d B S P L) 12.5 15.05.0 ISLE 10.7 Fourier Analysis in Audition 297 Chapter 10: The Auditory System a well-made violin. Well-made violins have a greater array of harmonics than do cheap violins. The relative loudness of different higher harmonics contributes to timbre, but there are other factors that contribute as well (see Chapter 12). To hear differences in timbre, go to ISLE 10.9. Phase The last characteristic of sound stimuli we consider is phase. Remember that sound is a change in sound pressure over time and space. Think about that clap again. It creates a wave of high-pressure peaks and low-pressure troughs that propagate across space. If you examine Figure 10.12, you see a sound wave with a peak at Time 1 (T1) and a trough at Time 2 (T2). Like any sound, we will hear this wave at the frequency that represents twice the distance from T1 to T2, that is, from one peak to the next one. Now consider Wave B. This wave has the same frequency as Wave A, but it is out of phase by 180 degrees. Thus, when the sound in Wave A is at its peak, Wave B is at its trough, and when the sound in Wave A is at its trough, the Wave B is at its peak. In Figure 10.12, we see what happens when we superimpose these sounds on each other: They cancel each other out. That is, when these sound waves are presented at the same time, but 180 degrees out of phase, we hear neither Wave A nor Wave B, as the two sounds cancel out. This is the principle behind noise-canceling headphones, which have become pop- ular among travelers. Some high-end cars also have noise-canceling systems to reduce the amount of road noise that one hears inside the car. You can see and hear a demonstration of phase and canceling in ISLE 10.10. TEST YOUR KNOWLEDGE 1. What is the fundamental frequency? Why is it important in sound perception? 2. What is the relation between timbre and harmonics? G4 G5 392 (Fundamental frequency) D6 G6 B7 D7 F7 G7 Frequency (Hz) A m pl itu de 0 0 1,000 2,000 3,000 4,000 5,000 Flute G4 G4 G5 392 (Fundamental frequency) D6 G6 B7 D7 F7 G7 Frequency (Hz) A m pl itu de 0 0 1,000 2,000 3,000 4,000 5,000 Violin G4 FIGURE 10.11 Fundamental Frequency and Harmonics of Common Musical Instruments The fundamental frequency is the lowest frequency produced when playing a particular note, though it need not be the loudest frequency. The higher harmonics provide overtones that make the sound richer and more complex; that is, they provide timbre. ISLE 10.9 Timbre and Overtones ISLE 10.8 Missing Fundamental Phase: the position in one cycle of a wave; there are 360 degrees in a single cycle of a wave ISLE 10.10 Phase and Cancellation FIGURE 10.12 Phase Phase refers to the position in one cycle of a wave. The two waves shown here have the same frequency but are 180 degrees out of phase with each other. If played together, they would cancel each other out, and we would hear nothing. Wave B Wave A Cancels sound 298 Sensation and Perception ANATOMY OF THE EAR 10.2 Describe the basic anatomy of the ear and the difference between the outer, middle, and inner ear. We now turn our attention to the anatomy of the ear. Like the eye, the ear is a complex system whose pur- pose it is to take energy from the external environment and render it into meaningful information. The ear fun- nels sound waves toward specialized hair cells in the inner ear that transduce the sound from physical sound energy into a neural impulse, which then travels to the auditory regions of the brain. Without transduction of the auditory stimulus, there would be no hearing. Thus, the question for this section is the following: How does transduction take place? In this section, we start with the entrance of the sound stimulus into the side of the head and follow it until transduction occurs. This pro- cess takes us through three distinct anatomical regions. These regions are the outer ear, the middle ear, and the inner ear. Figure 10.13 gives you an overview of the anatomy of the ear. The Outer Ear The part of your ear that sticks out on the side of your head is called the pinna (plural pinnae). The pinna collects sound and funnels it into the external auditory canal. The pinna’s fleshy shape helps gather sound waves and channel them to the ear. The shape of the pinna also helps in sound localization, that is, determining the direction a sound is coming from. In other animals, the pinna can be moved in various ways, helping these animals pinpoint sound in space (Figure 10.14). Think of your dog’s ear perking up as it hears someone’s car pulling into the driveway. This action helps the dog determine the source of the sound. A select few humans can also wiggle their pinnae, but it is unclear if this gives them an advantage in sound localization, although it may be fun at parties. After sound is collected by the pinna, it is directed into the external auditory canal. The external auditory canal (also known as the external auditory meatus) conducts sound from the pinna to the tympanic membrane. In all people, regardless of their height, the auditory canal is about 25 mm long (just shy of 1 inch). This length helps amplify certain higher frequencies. The auditory canal also acts to protect the tympanic membrane. The tympanic membrane is commonly known as the eardrum. The tympanic mem- brane is a thin elastic sheet that vibrates in response to sounds coming through the external auditory canal. Thus, sound moving down the auditory canal hits against the membrane, which vibrates in response to that sound. The tympanic membrane also seals the end of the outer ear. Damaging the tympanic membrane can result in hearing loss. A common injury in scuba diving is the puncturing of the tympanic membrane due to high pressure in the auditory canal. This injury is very painful but does not cause deafness. Divers may expe- rience temporary hearing loss until the membrane heals. In most cases, the tympanic membrane will repair itself. But repeated injury can result in permanent damage due to scarring, in which case hearing loss will be permanent. The Bajau people, a traditional diving and seafaring culture in the Philippines, will intentionally rupture their tympanic membranes as children to reduce pain from free diving (Langenheim, 2010). Not surpris- ingly, many Bajau people have hearing deficits. FIGURE 10.13 Overview of the Anatomy of the Ear Hammer Anvil Semicircular canals Cochlea Auditory nerve Stirrup Tympanic membrane (eardrum)External auditory canal Pinna Pinna: the structure that collects sound and funnels it into the auditory canal External auditory canal (external auditory meatus): the channel that conducts sound from the pinna to the tympanic membrane Tympanic membrane: a thin elastic sheet that vibrates in response to sounds coming through the external auditory canal; commonly known as the eardrum 299 Chapter 10: The Auditory System The Middle Ear The tympanic membrane is the last structure of the outer ear. The next part of the ear is called the middle ear, which consists of three small bones that transmit sound into the inner ear. When the tympanic membrane vibrates, it causes motion in these three small bones, called ossicles, which then conduct the sound mechanically. You have probably heard the old (but true) refrain that the ossicles are the smallest bones in the body, but we repeat it here anyway. The three ossicles act to amplify sound waves, although most of the amplification comes from the size of the tympanic membrane relative to the oval window. The tympanic membrane causes sound transmission in the first ossicle, known as the malleus. The malleus vibrates and transmits the sound to the next ossicle, known as the incus. The incus is then connected to the final ossicle, known as the stapes. The stapes then pushes against the oval window of the cochlea, and sound moves into the inner ear. Examine Figure 10.15 to see how these bones are connected inside the ear. For a simulation of how this system works to propagate sound, see ISLE 10.11. The ossicles are important for the following reason. The cochlea of the inner ear is immersed in liquid, but sound in the auditory canal is in the form of air-based sound waves. Why is this relevant? Sound travels well in water, but it does not cross from air to water very well. In fact, 90% of sound traveling through air will bounce off water rather than penetrate it. This is why that darn party across the lake seems so loud even though it is quite far away. In essence, sound just skips off the surface of the water rather than penetrating it. The sound that heads in the direction of the lake itself echoes off the lake’s surface rather than being absorbed by the water, leading it to sound much louder farther away than if the same party were being held at the edge of a field. In contrast, because sound is being transmitted from an air environment to the watery environment of the inner ear, the sound must be amplified as much as possible to overcome the loss of sound due to this transition. In the outer ear, the sound is still traveling through air as it makes its way to the tympanic membrane. Thus, a system is needed to amplify the sound before its intensity is diminished when it hits the ISLE 10.11 The Middle Ear Ossicles: three small bones in the middle ear Malleus: the first ossicle in the middle ear; receives vibrations from the tympanic membrane and transmits them to the incus Incus: an ossicle in the middle ear; receives vibrations from the malleus and transmits them to the stapes Stapes: an ossicle in the middle ear; receives vibrations from the incus and transmits them to the oval window of the inner ear FIGURE 10.14 Pinnae in Different Mammals © iS to ck ph ot o. co m /t et m c © iS to ck ph ot o. co m /m om ci lo g © iStockphoto.com /TakinPix © iS to ck ph ot o. co m /J iri H au re lju k © iS to ck ph ot o. co m /y es fo to © iStockphoto.com /Becart 300 Sensation and Perception liquid medium of the cochlea. The ossicles serve to amplify the sound. Interestingly, this is the reason why fish have no need for an outer and middle ear. Because sound goes directly from the water to their inner ears, they do not need to amplify the sound. The ossicles amplify sound in the following way. First, they use lever action to increase the amount of pressure change. A small amount of energy at the malleus becomes a larger amount at the stapes. Second, the ossicles transfer energy from a larger surface area, the tympanic membrane, to a smaller surface area, the oval window of the cochlea. Indeed, the tympanic membrane is about 18 times as large as the oval window. Thus, the bones increase the sound pressure 18 times at the oval window, critical in the transmission of sound into the liquid envi- ronment of the cochlea. Thus, most of the amplification is due to the dif- ference in size between the tympanic membrane and the oval window. The middle ear has a number of other important functions. One part of the middle ear is called the eustachian tube. The eustachian tube is the thin tube that connects the middle ear with the pharynx and serves to equalize air pressure on either side of the tympanic membrane. Normally, the eustachian tube is closed, but it opens briefly when we swallow or yawn, for example. The brief opening of the eustachian tube is the technical explanation for the phenom- enon of “popping” your ears. When you hold your nose and blow, the eustachian tubes open to release the pressure. This is also why divers must continually equalize as they descend to deeper depths (Figure 10.16). Failure to do so during diving brings first intense pain and also potential damage to the tympanic membrane. The middle ear is filled with air, and except for when the eustachian tube is open, it is cut off from any changes in air pressure in the out- side environment. For optimal operation of the tympanic membrane, it is important for the air pressure on both sides of the tympanic mem- brane to be equal, but the air pressure outside the tympanic membrane changes regularly due to weather and altitude. The eustachian tube opens to allow the two air pressures to equalize. Usually the change in air pressure in the middle ear is slight, which is why the eustachian tube does not open very often under normal circumstances. Yet when you ascend or descend a mountain, especially when you are doing so rapidly in a car, there can be a large change in air pressure, causing your ears to pop, particularly as you descend and the air pressure increases outside your ears relative to inside them. Pressure changes may occur when we dive in water, as pressure increases more rapidly as we descend underwater than it does in air. One of the first skills a scuba diver must master is “equalizing,” that is, forcibly opening the eustachian tubes to equalize pressure. If you are ascending a moun- tain, the air pressure outside your ears decreases with each step up in altitude. If your eustachian tubes stay closed, you will have much more air pressure on the middle-ear sides of your tympanic mem- branes than outside, causing them to bow toward the outside of your head. Being pushed out causes the tympanic membrane to move less effectively in response to sound stimuli, and in these cases you might find that your hearing is not quite as good as normal. When you yawn or swallow, in these cases there will be a rapid reduction of the air pressure inside your middle ear, which you experience as the popping of your ears. The ossicles also serve a role in attenuating sustained loud sounds. There is a muscle attached to the malleus called the tensor tympani and a second muscle attached to the FIGURE 10.15 Anatomy of the Middle Ear The three bones of the middle ear receive mechanical stimulation from the tympanic membrane and transmit it to the oval window. Because the oval window is smaller than the tympanic membrane, this transmission acts to amplify the sound. Eustachian tube: a thin tube that connects the middle ear with the pharynx and serves to equalize air pressure on either side of the eardrum Tensor tympani: the muscle that is attached to the malleus © iS to ck ph ot o. co m /P op ar tic FIGURE 10.16 The Middle Ear Adapts to Pressure Changes From Diving 301 Chapter 10: The Auditory System stapes called the stapedius. Like the bones they are attached to, these are the smallest mus- cles in the body. Their job is to tense in the presence of very loud noises, thus restricting the movements of the ossicles and avoiding damage to the inner ear. This acoustic reflex pro- tects somewhat against chronic loud noises, such as car stereos, but it is too slow for sud- den loud noises such as the sound of a gun firing. (Have we mentioned that being around gunfire is not good for your hearing?) Interestingly, these muscles may also tense in response to sounds generated inside the head, such as the sounds generated by chewing or talking. TEST YOUR KNOWLEDGE 1. Describe the anatomy of the outer and middle ear. 2. What are the functions of the ossicles? The Inner Ear The inner ear contains the parts of the ear that transduce sound into a neural signal. In par- ticular, the hair cells situated along the organ of Corti in the cochlea act by taking vibrations and converting them into a neural signal. In this way, the hair cells are equivalent to the rods and cones of the retinae. The inner ear is an amazing accomplishment of evolution, but it has lots of parts with hard-to-remember (and seemingly arbitrary) names, and thus it typically presents a challenge for students learning about it for the first time. Therefore, read carefully, study the diagrams, and test yourself repeatedly. For a diagram of the inner ear, examine Figure 10.17. Go to ISLE 10.12 for an interactive model of the workings of the inner ear. FIGURE 10.17 The Inner Ear The inner ear is a very complicated part of the body that contains the hair cells that transduce sound into a neural signal. These hair cells are located along the basilar membrane, which transmits the mechanical input of sound. When sound enters the inner ear, the tectorial membrane causes the basilar membrane to vibrate. These vibrations are detected by the hair cells, starting the neural signal. Hammer (a) Cochlea and middle ear (b) Section of cochlea (c) Organ of Corti Auditory nerve Vestibular canal Hair cells Tympanic canal Cochlear canal Tectorial membrane Hair cells Basilar membrane Cochlear neuron Oval window (membrane behind stirrup) Tympanic membrane Stirrup Anvil ammer (c) OrOrgangan of Coorti AudA itory nern ve Vestibulaarr canal HaiH r ccelle ss Tympanic canal CocC hlear canc al Tectorial membrane HaiHair celll ss Basilar membrane Cochlehl arr neun ronn Oval window (meembrmb aneane beb hini d sd tirrup) mmpanpa icic mbrane StiS rruup Stapedius: the muscle that is attached to the stapes Acoustic reflex: a reflex that tightens the tensor tympani and the stapedius in response to chronic loud noise Cochlea: the snail-shaped structure of the inner ear that houses the hair cells that transduce sound into a neural signal 302 Sensation and Perception The cochlea is the snail-shaped structure of the inner ear that houses the hair cells that transduce sound into a neural signal. The term cochlea derives from the Greek word for snail, and if you look at the cochlea depicted in Figure 10.17, you will see that its spiral indeed looks like a snail. The cochlea is a coiled tube, coiled 2.74 times. As a coiled tube, it takes up just about 4 mm of space inside your ear, but if you were to unroll it, it would stretch to about 33 to 35 mm in length (1.3 inches). Now examine the picture of the cross-section of the cochlea, shown in Figure 10.18. You will see that the cochlea has three liquid-filled chambers, called the tympanic canal, the mid- dle canal (also called the cochlear duct), and the vestibular canal. In the apex (the end of the cochlea) is an opening called the helicotrema that allows fluid to flow between the tym- panic canal and the vestibular canal. The apex is the part of the cochlea farthest from the oval window if the cochlea were uncoiled. Seated just underneath the oval window is the round window. The round window is a soft tissue substance at the base of the tympanic canal. Its function is as an “escape” valve for pressure from sounds that arrive in the cochlea, because liq- uids do not behave the way that gasses do. Remember that in sound waves, the molecules com- press, but in liquids, such as those that fill the cochlea, the molecules compress much less, so when the oval window presses in, there needs to be a place that pushes out, namely, the round window. We will return to function shortly, but we need to cover just a bit more anatomy first. The Basilar Membrane of the Cochlea Two membranes separate the canals of the cochlea. One is Reissner’s membrane, which separates the vestibular canal and the middle canal. The other membrane is the basilar 2000 Hz 1000 Hz 800 Hz 400 Hz 600 Hz 3000 Hz 5000 Hz 4000 Hz 10000 Hz 20000 Hz FIGURE 10.18 Cross-Section of the Cochlea The cochlea is the snail- shaped structure. The base is the starting point and then it coils around to the apex. The basilar membrane separates the middle canal from the tympanic canal. The organ of Corti rests along the basilar membrane. Reissner’s membrane separates the vestibular canal from the middle canal. Tympanic canal: one of the three chambers in the cochlea; separated from the middle canal by the basilar membrane Middle canal (cochlear duct): one of the three chambers in the cochlea; separated from the tympanic canal by the basilar membrane; contains the organ of Corti Vestibular canal: one of the three chambers in the cochlea; separated from the middle canal by Reissner’s membrane Round window: a soft tissue substance at the base of the tympanic canal whose function is as an “escape” valve for excess pressure from loud sounds that arrive in the cochlea ISLE 10.12 The Basilar Membrane and Sound Stimuli 303 Chapter 10: The Auditory System membrane, which separates the tympanic canal and the middle canal. Lying along the basilar membrane within the middle canal is the organ of Corti, which contains the hair cells. The basi- lar membrane is not really a membrane. Rather, it is composed of harder and thicker material, which allows it to vibrate in response to incom- ing sound. At the base (nearest to the oval and round windows), the basilar membrane is thicker and stiffer and therefore more respon- sive to high-frequency sounds. At the apex (the uncoiled opposite end), the basilar membrane is less thick and less stiff and therefore more responsive to low-frequency sounds. Take a look at Figure 10.19. Notice that the stapes of the middle ear asserts pressure on the oval window. This pressure from the stapes causes a wave in the fluid (perilymph) of the middle canal in the inner ear. This pres- sure wave in the perilymph causes a traveling wave to move down the length of the basilar membrane. Traveling wave here means that the wave moves from the base to the apex of the basilar membrane (ISLE 10.13). If the sound being presented is high frequency, the traveling wave will show greater motion toward the base of the basilar membrane. Lower frequency sound will cause more movement farther down the basilar membrane. The lowest frequency sounds we can hear move the basilar membrane near the apex. The movement of this wave is critical to understanding how frequency is coded in the cochlea. To understand why the thick end of the basilar membrane is conducive to high-fre- quency sounds and the thinner or “floppy” end to lower frequency sounds, think of stringed instruments. On a violin, for example, lower pitches are produced by the G string, located farthest to the left on the fingerboard. Pull on the G string, and it will feel relatively loose. The highest pitches on a violin are produced by the tightly wound E string, which is farthest to the right on the fingerboard. Pull on this string, and it feels tight. If the E string is out of tune and too flat (too low in pitch), the violinist gets it into tune by tightening the string. Because the E string is so tightly wound to begin with, many novice violinists may snap the string altogether. If the E string is out of tune and too sharp (too high in pitch), the violinist gets it into tune by loosening the string. The basilar membrane is equivalent to this. The base is tightly wound and thus more responsive to high frequencies, whereas the apex is loose and responsive to lower frequencies. For an illustration of how the basilar membrane moves in response to dif- ferent frequencies, go to ISLE 10.14. Incoming vibrations will cause the basilar membrane to vibrate. The base of the bas- ilar membrane is tightly wound and thus more responsive to high frequencies, whereas the apex of the basilar membrane is loose and responsive to lower frequencies. Indeed, the displacement of the basilar membrane in response to frequency is quite specific. Each location along the basilar membrane responds to a characteristic frequency. Any sound will move the basilar membrane at every location, but a par- ticular location will respond the most, in terms of movement, to its characteristic frequency. In the most common case, in which we hear a complex sound, complete with harmonics, the basilar membrane will reflect these frequencies by moving at various places along its length. In effect, the basilar membrane does a Fourier analysis, breaking down complex sounds into their component frequencies (von Békésy, 1960). To see this visually, look at Figure 10.20. You can also see a demonstration of this on ISLE 10.15. Reissner’s membrane: the membrane that separates the vestibular and middle canals Basilar membrane: the membrane that separates the tympanic canal from the middle canal; the organ of Corti lies on the basilar membrane Organ of Corti: a structure on the basilar membrane that houses the hair cells that transduce sound into a neural signal Perilymph: the fluid that fills the tympanic canal and the vestibular canal Characteristic frequency: the frequency to which any particular location along the basilar membrane responds best Perilymph Basilar membrane At thicker, narrower, stiffer base of basilar membrane, high-frequency pressure waves cause greatest displacement. In middle portion of basilar membrane, mid-frequency pressure waves cause greatest displacement. At thinner, wider, more flexible apex of basilar membrane, low-frequency pressure waves cause greatest displacement. Vibration of stapes against oval window causes pressure wave in perilymph. Pressure wave in perilymph causes traveling wave in basilar membrane. FIGURE 10.19 Sound Waves and Changes to the Basilar Membrane ISLE 10.13 The Travelling Wave ISLE 10.14 Place Code Theory ISLE 10.15 The Basilar Membrane and Fourier Analysis 304 Sensation and Perception FIGURE 10.20 Frequency and Displacement Along the Basilar Membrane (a) The basilar membrane is shown as if it were uncoiled and spread in a line. Each location along the basilar membrane vibrates to the greatest extent in response to one frequency. That is, a sound of a particular frequency will maximally displace the basilar membrane at that location. (b) We can see characteristic frequencies and their displacement along a schematic basilar membrane. Sounds of high frequency excite locations toward the base of the basilar membrane, whereas sounds of low frequency excite locations toward the apex of the basilar membrane. Apex t 0 t 1 t 2 t 3 t4 t 0 t 1 t 2 t 3 t4 Base Time Wave responses to sound dominated by high frequencies Traveling waves in basilar membrane (displacement of basilar membrane from base to apex over time, from t 0 to t 4 ) 20,000 Hz 1,000 Hz 20 Hz 10,000 Hz 100 Hz Base Characteristic frequencies in basilar membrane Apex Time Wave responses to sound dominated by low frequencies (a) (b) Hair cells: cells that have stereocilia for transducing the movement of the basilar membrane into a neural signal Stereocilia: the hairlike parts of the hair cells on the top of the inner and outer hair cells Outer hair cells: cells that sharpen and amplify the responses of the inner hair cells Inner hair cells: cells that are responsible for transducing the neural signal Tectorial membrane: a membrane that rests above the hair cells within the organ of Corti In essence, the basilar membrane converts the sound energy that beats against the oval window into mechanical movement along its length with any particular location along its length responding to particular frequencies of sound. We are now, finally, ready to begin to discuss how this physical signal gets changed into a neural one. The Organ of Corti The organ of Corti is the structure along the basilar membrane that contains the hair cells that transduce sound into a neural signal. The name Corti honors the Italian sci- entist Alfonso Corti (1822–1876), who experimented on the function of the cochlea in the 1850s. In addition to the hair cells, the organ of Corti also contains the dendrites of the auditory nerve that brings the neural signal to the brain. The critical hair cells in the organ of Corti are specialized cells for transducing the motion of the basilar membrane into a neural signal. The hair cells have hairlike filaments called stereocilia for transducing the movement of the basilar membrane into a neural signal. These ste- reocilia bend in response to the movement of the basilar membrane, and the bending changes the voltage within the hair cells. There are four layers of hair cells that follow the basilar membrane. The first row is called the outer hair cells, and the row that lies inside that is called the inner hair cells (Figure 10.21). There are about 3,500 inner hair cells and about 3 times that many outer hair cells. Each hair cell has anywhere from 50 to 150 stereocilia sticking off its top in order to detect the motion of the basilar mem- brane. Inner hair cells are responsible for transducing the neural signal, whereas outer hair cells refine and amplify the neural responses of the inner hair cells. The tectorial membrane sits above the hair cells within the organ of Corti, helping hold them in place (ISLE 10.16). ISLE 10.16 Transduction and Hair Cells 305 Chapter 10: The Auditory System When a vibration of the basilar membrane causes the basilar membrane to move upward, the stereocilia brush against the tectorial membrane. When stereocilia are pushed in this manner, there is a change in the cell’s voltage potential. This voltage change causes the release of neu- rotransmitters, which cause the auditory nerve to send a signal. This process can be seen visually in Figure 10.22. And with the induc- tion of a signal in the auditory nerve, the sound has been trans- duced and sound information gets conveyed to the brain. One of the most impor- tant scientists in this area was Hungarian-born Georg von Békésy (1899–1972) (Figure 10.23). He did most of his important work while on the faculty of Harvard University. For his work on the basilar membrane and how it codes for frequency, von Békésy won the 1961 Nobel Prize in Medicine and Physiology. Von Békésy was the first to show that the basilar membrane vibrates more at certain locations along its length in response to different frequencies, thus allowing the auditory system to discrim- inate among frequencies. Although many of you may find this bizarre, von Békésy did his most famous work by playing really loud sounds into the ears of human cadavers. Much of the discussion of basilar membrane function in the previous paragraphs is knowledge we learned from von Békésy’s work. His view of frequency representation in the cochlea is known as the place code theory. The alternative view is known as the temporal code theory (Wever & Bray, 1937), which states that frequency represen- tation occurs because of a match between sound frequency and the firing rates of the auditory nerve. The evidence for place code theory rests largely on the observation that the bas- ilar membrane’s thickness changes along the cochlea, and this thickness allows it to FIGURE 10.21 The Organ of Corti The organ of Corti contains both inner hair cells and outer hair cells, as well as the tectorial membrane. The outer hair cells connect to the auditory nerve. Auditory nerve Type I nerve fiber Endolymph Tectorial membrane Inner hair cell Type II nerve fiber Movements of basilar and tectorial membranes, resulting in shearing force on stereocilia Stereocilia Basilar membrane Outer hair cell Place code theory: the view that different locations along the basilar membrane respond to different frequencies Temporal code theory: the view that frequency representation occurs because of a match between sound frequency and the firing rates of the auditory nerve Outer hair cellsInner hair cells Tectorial membrane Basilar membrane Upward phase Shear force Downward phase Shear force Sound-induced vibration Resting position FIGURE 10.22 Actions of Hair Cells Vibration causes the tectorial membrane to shear against the hair cells, that is, to move in opposite directions. This movement causes the stereocilia of the hair cells to transduce sound into a neural signal. 306 Sensation and Perception respond differently to different frequencies along its length. Von Békésy also observed the movement of the basilar membrane in response to loud sounds. In this way, the basilar membrane separates out different frequencies, and the hair cells that lie along the length of the basilar membrane can code for these different frequencies. Thus, when hair cells at a specific position are activated, the auditory system can interpret this as a sound at a particular frequency. This is illustrated in Figure 10.20 as well as in ISLE 10.17. As we get older, the basilar membrane gets stiffer. This accounts for the observa- tion that as we age, we lose the ability to hear higher frequencies. By the time people reach the age of 50 or so, they cannot hear frequencies above approximately 12,000 Hz. This occurs even in the absence of exposure to loud sounds. Even people at 25, by and large, have lost hearing above 16,000 Hz. This age-related hearing loss in the higher frequency range has some amusing consequences. Teenagers can purchase ring tones for their cell phones at very high frequencies that will alert them to incom- ing messages, but to which their teachers will be deaf. Most adults are not simply impaired when it comes to hearing a 20,000 Hz tone; they are deaf. Thus, a teenager receiving a call or text message has little risk of being detected by anyone else than another teenager. On the other hand, some store owners, in an effort to keep teenagers from “hanging out” in front of their stores, will play very loud high-frequency tones. The store owners and their paying customers are deaf to these sounds and will there- fore come and go as they like. But the loud high-frequency sounds will be painful to the loitering teens, who will therefore seek somewhere else to hang out. Tit for tat, we suppose. TEST YOUR KNOWLEDGE 1. Describe how hair cells transduce the sound signal in the cochlea. 2. What is place code theory? How does the displacement of the basilar membrane in response to specific frequencies support this view? 10.3 Examine the nature of hearing loss and what can be done to improve hearing in hearing-impaired individuals. Sc ie nc e H is to ry Im ag es /A la m y St oc k Ph ot o FIGURE 10.23 Georg von Békésy ISLE 10.17 Temporal Code Theory EXPLORATION: Hearing Loss Those who have lost the ability to hear can learn language through sign languages and may often be quite competent at spoken languages by reading lips and inferring contexts (Figure 10.24). Many deaf people have accomplished great things even without this crucial sensory system. In a beautiful book, Josh Swiller (2007), a profoundly hear- ing-impaired man, describes his adventures as a Peace Corps volunteer in southern Africa. In his tireless efforts to help the people of his assigned village, Swiller does not let his hearing impairment hold him back. Nonetheless, when his digital hearing aids are stolen while he is thou- sands of miles away from possible replacements, Swiller begins to understand what it is like to be truly deaf in a world in which hearing is necessary. We have already mentioned a number of ways in which the auditory system is vulnerable to damage. In particular, the mechanisms of the ear, while finely tuned to an incredible range of amplitudes and frequencies, are also vulnerable to loud noises and high pressures. And even though each hair cell has a great many stereocilia, there are only so many hair cells, so that damage to hair cells can have grave con- sequences for hearing also. Luckily, there are many options to improve hearing in those with hearing loss. In the United States, more than 30 million people have hear- ing loss, according to the Centers for Disease Control and Prevention (2017). Approximately three quarters of these people are over the age of 60. Presbycusis refers to the loss 307 Chapter 10: The Auditory System of hearing associated with aging (Yost, 2007), which is more common in men than women. Presbycusis is also more com- mon in societies such as our own, in which more people are exposed to loud sounds than people are in more traditional societies, where they may not be subjected to all the mechan- ical sounds people in industrialized societies are. (Think of how many mornings you have awoken to the sounds of lawn care machines outside your window—and now think of the landscapers’ hearing.) There are also a number of genetic and acquired hearing loss problems in younger adults. Disease and injury can also cause hearing loss in people of all ages. Audiologists are doctoral-level specialists who evaluate, diag- nose, and treat hearing impairments, provided they do not require immediate medical attention. Audiologists are equiva- lent to optometrists in the visual domain, in that their specialty is assessing the problem and then helping the patient find the right hearing aid, just as an optometrist evaluates visual prob- lems and then prescribes corrective eyewear. During an audi- ological exam, an audiologist administers a set of tests using an audiometer. It may be that you have had an audiological exam at some point. The audiometer presents pure tones at set frequencies and known amplitudes to either the left or right ear. It is used in the assessment of absolute threshold at each frequency for a patient. The result of this procedure is an audiogram. Figure 10.25 shows an audiological report of a patient with a genetic sensorineural hearing impairment (i.e., hearing loss due to congenital damage to hair cells). This patient’s peak impairment is around 4,000 Hz in both ears. The patient has mostly normal hearing at lower frequencies. We now consider the various forms of hearing impairment. Conductive Hearing Loss Conductive hearing loss is characterized by damage to some aspect of sound transmission in the outer or middle ear. Thus, in conductive hearing loss, sound does not properly get to the cochlea for transduction into a neural signal. This can occur because of blockage of the auditory canal, a torn tympanic membrane, or damage to the ossicles. In particu- lar, conductive hearing loss may occur because of otoscle- rosis. Otosclerosis is an inherited bone disease in which the ossicles, particularly the stapes, may calcify and therefore be less conductive of sound. In some cases, hearing loss due to otosclerosis can be improved by replacing the stapes with an artificial bone. There are also conditions in which hearing loss may occur on a temporary basis. Ruptured tympanic membranes may heal properly, restoring full hearing. Hearing loss may also occur on a temporary basis during an ear infection, if the auditory canal becomes blocked with fluid. Treatment of the infection may be sufficient to return hearing to normal, but in some cases, the middle ear must be drained to allow hearing to return to normal. Sensorineural Hearing Loss Sensorineural hearing loss occurs because of damage to the cochlea, the auditory nerve, or the primary auditory cortex. The term sensorineural hearing loss refers to acquired © iStockphoto.com /D ean M itchell FIGURE 10.24 Sign Language Normal impairment Caused by noise exposure Left ear Right ear 125 −10 10 20 30 40 50 60 70 80 0 250 500 FREQUENCY in Hz H e a ri n g l e v e l in d B 1000 2000 4000 8000 FIGURE 10.25 Audiological Report This patient has moderate to severe hearing loss in both ears in the high frequency range. But you can also see that some frequencies are more affected than others. Source: Reprinted with permission from the Handbook for Acoustic Ecology, B. Truax, ed., Cambridge Street Publications, CD-ROM edition, 1999. Otosclerosis: an inherited bone disease in which the ossicles, particularly the stapes, may calcify and therefore be less conductive of sound 308 Sensation and Perception hearing problems and genetic problems, and the condition can range from minor hearing loss to profound hearing deficits. This form of hearing loss is most often precipi- tated by damage to hair cells, which can occur because of noise exposure or certain drugs. In addition, certain anti- biotics and cancer treatments have the side effect of dam- aging hair cells in the cochlea. Although these medicines are given only when a patient’s life is at stake, they can affect hearing. Inherited sensorineural hearing loss is rare but may show up at infancy, in childhood, and sometimes not until adulthood. Today, tests are immediately given to a newborn’s ear to determine if there is congenital hearing loss. If so, parents can make appropriate decisions for their infant in order to make sure that the infant develops lan- guage normally. Tinnitus Tinnitus is the condition in which people perceive sounds even when none are present. The perceived sound usually sounds like a pure tone at a particular frequency. Most peo- ple experience tinnitus occasionally and for brief periods of time, and for the vast majority of us, tinnitus is seldom a problem. But when it occurs all the time or is subjectively loud, it becomes a problem (Figure 10.26). It is thought that most cases of tinnitus involve a neural signal being sent to the brain in the absence of an actual sound. Thus, tinni- tus can interfere with the perception of real sounds in the environment, including speech. Tinnitus is associated with noise-induced hearing loss, but there may be other causes as well (Møller, 2006). A former student of Dr. Schwartz was afflicted with a bad case of tinnitus after being awakened by a loud fire alarm just above his head. The illusory sound was so loud for this young man that it interfered with his ability to study. After several months, the tinnitus decreased and eventually disappeared. In other cases, tinnitus may be the result of damage to the cochlea or damage or infection to the auditory nerve. Treatment may include hearing aids, which can amplify external sounds, such as the frequencies of human voices, while at the same time playing white noise at the same frequency as the tinnitus sound, in hopes of canceling it out. Some attempts to find drugs that reduce tinnitus have also been successful. Tinnitus: a condition in which people perceive sounds even when none are present © iS to ck ph ot o. co m /a po m ar es FIGURE 10.26 Tinnitus APPLICATION: The Science of Hearing Aids and Cochlear Implants Hearing Aids Hearing aids are electronic devices that amplify sound so that people with hearing deficits can hear sounds that oth- erwise would be below their thresholds. Hearing aids can be the “behind-the-ear” style (see Figure 10.27a) or the “in-the-ear” style (Figure 10.27b). In-the-ear hearing aids are smaller and less obtrusive than behind-the-ear aids, but behind-the-ear devices can amplify sounds more and thus are preferable for people with moderate, severe, or profound hearing loss. In-the-ear aids work fine for people with mild to moderate hearing loss. Younger people with hearing loss may also prefer in-the-ear devices because they are less visible. In the past, hearing aids were analog devices. Analog devices amplify the actual sound present, but because they work with the actual sound, they are limited in what other forms of sound processing they can do. In general, analog hearing aids amplify all sounds the same amount (though Hearing aids: electronic devices that amplify sound so that people with hearing deficits can hear sounds that otherwise would be below their thresholds 309 Chapter 10: The Auditory System many have volume controls). This is potentially problem- atic because, as we have just seen, the pattern of hear- ing loss may be different at different frequencies. Thus, if we amplify all sounds, we may amplify some to a level at which they can be heard, but that may mean that other sounds are too loud and others are still too soft. In technologically advanced countries, most hearing aids are now digital devices. Digital hearing aids first con- vert the sound signal into a computer code, which is then reconverted into an analog sound for the wearer. This transformation into a digital signal allows additional processing to the signal in addition to amplification. Digital hearing aids have a number of important advan- tages over earlier analog aids. First, digital hearing aids can be fitted to a particular individual’s pattern of hear- ing loss. This is important because hearing loss may vary greatly from one patient to another, both in the loudness of sounds required to allow hearing and in the frequency range of the hearing loss. Digital hearing aids can be pro- grammed to amplify some frequencies more than other frequencies. If a patient has nearly normal hearing at 400 Hz, a digital hearing aid can be programmed to amplify not at all or very little at that frequency. But if the same patient has severe hearing loss at 1,000 Hz, the aid can be programmed to amplify much more at that frequency. In this way, the digital aid can restore a pattern of hear- ing that approximates what a person with normal hear- ing would experience. Digital hearing aids can also be programmed to have directionally sensitive microphones. With this feature, a person with hearing loss can amplify sound in front of herself, allowing better hearing of a conversation, and, at the same time, block out irrelevant sounds coming from other directions. This may be useful while trying to have conversations in noisy environments such as in restau- rants or cars. In another situation, in which a person does not know the direction from which relevant sounds will be coming, the directional microphone can be switched off. Digital hearing aids can also be programmed to amplify different frequencies at different times. People with digital hearing aids can adjust the programs they are running to suit their needs. For example, a first-grade teacher may have a program in her hearing aid that amplifies the fre- quencies likely to belong to first graders’ voices while she is teaching. She may turn that program off during recess, when the kids are running around on the playground. Later, when she gets home, she can switch to a program that allows her to amplify the deeper voices of adult male relatives. Hearing aids may also be obtained that have external microphones that can be placed near the desired sound source, and the sound will be electronically transmitted wirelessly into the hearing aid. Consider a student with a hearing deficit. He can request that his professor wear a microphone while teaching the class. The professor need only place the microphone around her neck and speak normally. The sound will be wirelessly sent to the hear- ing aid, and the student will hear the professor’s voice directly from the aid rather than across the room. This aspect of hearing aids can be beneficial in restaurants as well, as the microphone can be placed around the neck of the person being spoken to. Then, despite the noise of the background, the voice will be transmitted directly into the receiver of the hearing aid. Some hearing aids also come equipped with similar systems for watching television. The audio transmission of the television can be transmit- ted directly into the hearing aid (e.g., Bluetooth systems). Such equipment, however, is often expensive and may not be covered by insurance. Thus, many people with insuffi- cient insurance or ability to afford aids may have to settle for less complete systems. In the United States, hearing aids are covered by Medicare, so older adults should have access to quality aids. FIGURE 10.27 Hearing Aids (a) Behind-the-ear aids. (b) In-the-ear aids. Hearing aids amplify sounds so that the person can hear better. Modern digital hearing aids can be adjusted to amplify only those frequencies at which a person is impaired. © iS to ck ph ot o. co m /A va ila bl eL ig ht © iS to ck ph ot o. co m /a er og on do (a) (b) 310 Sensation and Perception Hearing aids require the presence of a minimum of some transmission of a sensorineural signal from the hair cells to the auditory nerve. Hearing aids amplify sound, allow- ing damaged ears to hear sounds, but they do not replace the need for a functioning cochlea. However, there is another tool that can be used to restore hearing in those who have become clinically deaf because of damage to the cochlea (sensorineural deafness). This is the domain of the cochlear implant, in which a mechanical device essentially replaces the hair cells along the basilar membrane. We turn next to cochlear implants. Cochlear Implants Cochlear implants are designed to restore some hear- ing, typically of spoken voices, to deaf individuals. Cochlear implants stimulate the auditory nerve artifi- cially with an electronic system, replacing the hair cells of the cochlea in the patient receiving the implant. This restores some hearing to patients who have become deaf. Some parents of deaf children are also electing to have cochlear implants placed in their deaf chil- dren’s inner ears to allow them to develop the ability to understand spoken language. To be eligible for the procedure, one must have profound hearing loss or deafness in both ears, the hearing loss must be due to sensorineural problems, but the auditory nerve must be intact. Until recently, the procedure was done in just one ear, but having implants in both ears is now becoming more common. As of 2016, it was estimated that there were more than 300,000 people with cochlear implants worldwide (National Institute on Deafness and Other Communication Disorders, 2016). To place a cochlear implant into someone’s ear requires minor surgery to place components inside the ear. Then it requires anywhere from a few weeks to several months of training and therapy to allow the person to be able to inter- pret the sounds he is now hearing again (or, in children, for the first time). The cochlear implant works by having both external (outside of the ear) components and internal (inside the cochlea) components. The cochlear implant sys- tem is illustrated in Figure 10.28. The first external component is a miniature microphone. The microphone picks up sounds from the environment. The microphone is attached to a small processor, which is essen- tially a minicomputer. The processor filters out noise and sounds not in the speech range, so that the patient can focus on hearing language-related input. It can be reprogrammed for other purposes as well, such as a person who has gone deaf but wants to experience music again. The processor sends a signal along a small wire to a transmitter. The trans- mitter sits directly on the surface of the skin, attached to the bone just behind the ear. During surgery, a small magnet is placed on the inside of the skull to keep the transmitter in place. This magnet is barely noticeable but acts to keep the transmitter firmly in place when the implant is being used. The transmitter sends a signal to the internal components of the cochlear implant wirelessly via radio waves. Underneath the skin are two components that are per- manently in place after surgery. First, there is a receiver/ stimulator that picks up the radio waves from the external device. The receiver/stimulator converts the radio waves into an electric signal, which then travels by wire into the cochlea. Within the cochlea, set along the length of the basilar membrane, are a series of electrodes, perhaps as many as 30. (In the early days of cochlear implants, there may have been only 4 electrodes, but now up to 30 can be safely placed inside the cochlea.) These electrodes, when stimulated by the receiver/stimulator, will induce a neural signal in the auditory nerve fibers that normally would be stimulated by hair cells. The electrodes are placed along the length of the basilar membrane, so when an electrode fires, the person will hear a particular frequency associ- ated with that location on the basilar membrane. Thus, the cochlear implant uses an electric system to induce a normal neural signal in the auditory nerve. Because thousands of hair cells are being replaced by only a couple dozen electrodes, patients should not expect, Cochlear implants: devices that are designed to restore some hearing, typically of spoken voices, to deaf individuals; they stimulate the auditory nerve artificially with an electronic system, replacing the hair cells of the cochlea FIGURE 10.28 Cochlear Implants Cochlear implants allow deaf people to hear again by electrodes placed inside the cochlea. These electrodes are connected to an external microphone, which selectively stimulates electrodes depending on what frequencies are present in the sound signal. The electrodes then directly stimulate auditory nerve fibers. Transmitter (on the surface of skin) Receiver/stimulator (under skin) Cochlea EardrumWire to electrode in cochlea Microphone (with speech processor behind ear) 311 Chapter 10: The Auditory System nor will they receive, their old hearing back. Sounds will not be perceived as richly with an implant as they were before the person went deaf. Moreover, most patients will not immediately be able to understand speech again once the system is online. Patients will require many hours of training and therapy to basically relearn to hear with their new implants. This may be frustrating at first, but eventu- ally, almost all patients learn to interpret their new hear- ing abilities. Once this training occurs, many patients can regain their understanding of speech at rather high levels. And some patients may find the sound quality sufficient to enjoy listening to music again. One of the more controversial issues with respect to cochlear implants is whether or not cochlear implants should be placed in infants who are born congenitally deaf. Many in the sign language deaf community object to this practice, as they argue that such children can develop normally and use language appropriately if they start with sign languages from an early age. They argue that forcing these children into the hearing world will actually be a handicap in and of itself, when they could develop language normally through visual sign languages. Nonetheless, many parents, particularly hearing parents, elect to go with cochlear implants. If cochlear implants are placed in an infant’s ears within the first year of life, with proper commitment on the part of the parents and involved professionals, that infant is likely to grow up being able to understand speech and to develop normal speech. In some cases, the child may even be able to partic- ipate in music activities later on. As you can see in Figure 10.29, the external components of a cochlear implants are not large and look similar to an in-the-ear hearing aid, with the addition of the transmitter. They will not impede a young person from doing other activities. FIGURE 10.29 Cochlear Implants A child born deaf but equipped with cochlear implants can learn to play and appreciate music. © A gencja Fotograficzna Caro/A lam y CHAPTER SUMMARY 10.1 Discuss the physical nature of sound and how it correlates with the perceptual nature of sound. The sound stimulus consists of the periodic variations in air pressure traveling out from the source of the variations. Sound waves are the waves of pressure changes that occur in the air as a function of the vibration of the source. A cycle is the amount of time between one peak of high pressure and the next. A pure tone is a sound wave in which changes in air pressure follow a sine wave pattern. Complex tones, however, have many different pure tones mixed together, making their waveforms more complex. The amplitude of a sound is expressed as the difference between its maximum and minimum sound pressure. Amplitude is measured on the decibel (dB) scale. Loudness is the perceptual experience of amplitude or the intensity of a sound stimulus. Persistent exposure to sounds over 85 dB can result in hearing loss, and even a one-time exposure to sounds over 135 dB can result in damage to hearing. The frequency of a sound stimulus is the number of cycles that occur in 1 second. Frequency is measured in hertz (Hz). Pitch is the subjective experience of sound that is most closely associated with the frequency of a sound stimulus, related to the experience of whether the sound is high or low (such as the two ends of the keyboard of a piano). Humans can hear frequencies as low as 20 Hz and, depending on age, as high as 20,000 Hz. As we get older, we lose hearing in the high-frequency range. Harmonics are higher frequencies present in a complex sound that are inte- ger multiples of the fundamental frequency. The fundamental frequency is the lowest frequency in a complex sound and the one that determines the perceived pitch of that sound. Timbre is the musical term that refers to the perceived sound differences between sounds with the same pitch but pos- sessing different higher harmonics. Phase is the position in one cycle of a wave. There are 360 degrees in a single cycle of a wave. Sensation and Perception312 10.2 Describe the basic anatomy of the ear and the dif- ference between the outer, middle, and inner ear. The outside portion of the ear is called the pinna, and it col- lects sound and funnels it into the external auditory canal. The external auditory canal conducts sound from the pinna to the tympanic membrane. The tympanic membrane, com- monly known as the eardrum, is a thin elastic sheet that vibrates in response to sounds coming through the external auditory canal. The tympanic membrane causes changes in the middle ear. The sound is transmitted by three small bones called ossicles (the malleus, incus, and stapes). The malleus receives vibrations from the tympanic membrane and transmits them to the incus. The incus receives vibra- tions from the malleus and transmits them to the stapes. The stapes receives vibrations from the incus and transmits them to the oval window of the inner ear. The eustachian tube is a thin tube that connects the middle ear with the pharynx and serves to equalize air pressure on either side of the eardrum. The stapes beats against the oval window of the inner ear. The cochlea is the snail-shaped structure of the inner ear that houses the hair cells that transduce sound into a neural signal. The basilar membrane is the band of fibers that separate the tympanic canal from the middle canal. The organ of Corti lies on the basilar mem- brane. The organ of Corti is a structure that houses the hair cells that transduce sound into a neural signal. Inner hair cells have stereocilia for transducing the movement of the basilar membrane into a neural signal. Place code theory is the view that different locations along the basilar mem- brane respond to different frequencies. 10.3 Examine the nature of hearing loss and what can be done to improve hearing in hearing-impaired individuals. Hearing loss affects millions of people. Presbycusis is the loss of hearing associated with aging. An audiometer presents pure tones at set frequencies and known amplitudes to either the left or right ear. An audiogram is a graphical display of the audi- tory sensitivity of a patient compared with a standard listener. Conductive hearing loss occurs because of structural damage to the outer or inner ear. Otosclerosis is an inherited bone dis- ease in which the ossicles, particularly the stapes, may calcify and therefore be less conductive. Sensorineural hearing loss occurs because of damage to the cochlea, the auditory nerve, or the primary auditory cortex. Tinnitus is the condition in which people perceive sounds even when none are present. Hearing aids are electronic devices that amplify sound so that people with hearing deficits can hear sounds that otherwise would be below their thresholds. Digital hearing aids provide a set of fea- tures that can allow a person with hearing loss to function well in the hearing world. Cochlear implants are designed to restore some hearing, typically of spoken voices, to deaf individuals. Cochlear implants stimulate the auditory nerve artificially with an electronic system, replacing the hair cells of the cochlea. REVIEW QUESTIONS 1. What is sound? How is it transmitted through the environment? How is it measured? 2. What are the relations between amplitude and loudness, frequency and pitch, and waveform and timbre? 3. Why do loud sounds potentially cause damage to hearing? How loud must sound be to cause dam- age if heard repeatedly? How loud must sound be to cause immediate damage? 4. What are harmonics? How do they relate to the fun- damental frequency? What do we hear when the fundamental frequency is deleted from a sound? 5. Describe the anatomy of the outer and middle ear. How does sound get transmitted from the pinna to the oval window? 6. What are the ossicles? How do they transmit sound, and how do they amplify sound? Why is their amplifi- cation needed for normal hearing? 7. Describe the anatomy of the inner ear. How does sound get from the oval window to the hair cells of the organ of Corti? 8. What is place code theory? How does it describe the relation of place along the basilar membrane to our perception of pitch? 9. What is the difference between conductive hearing loss and sensorineural hearing loss? How does loud noise cause hearing loss? 10. What is the difference between hearing aids and cochlear implants? How does each improve hearing in the hearing impaired? Chapter 10: The Auditory System 313 PONDER FURTHER 1. How would our hearing be different if we could hear lower frequencies, as do elephants, or higher fre- quencies, as do dogs? 2. Describe the different ways in which sound travels through the ear, from the outer ear to the auditory nerve. KEY TERMS Acoustic reflex, 301 Amplitude, 292 Basilar membrane, 302 Characteristic frequency, 303 Cochlea, 302 Cochlear implants, 310 Complex sound, 296 Cycle, 291 Decibel (dB), 293 Eustachian tube, 300 External auditory canal (external au- ditory meatus), 298 Fourier analysis, 296 Frequency (sound stimulus), 294 Fundamental frequency, 296 Hair cells, 304 Harmonics, 296 Hearing aids, 308 Hertz (Hz), 294 Incus, 299 Inner hair cells, 304 Loudness, 293 Malleus, 299 Middle canal (cochlear duct), 302 Organ of Corti, 303 Ossicles, 299 Otosclerosis, 307 Outer hair cells, 304 Perilymph, 303 Phase, 297 Pinna, 298 Pitch, 294 Place code theory, 305 Pure tone, 292 Reissner’s membrane, 302 Round window, 302 Sound stimulus, 290 Sound waves, 291 Stapedius, 301 Stapes, 299 Stereocilia, 304 Tectorial membrane, 304 Temporal code theory, 305 Tensor tympani, 300 Timbre, 296 Tinnitus, 308 Tympanic canal, 302 Tympanic membrane, 298 Vestibular canal, 302 Sensation and Perception314 Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review, study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning. Learning Objectives Digital Resources 10.1 Discuss the physical nature of sound and how it correlates with the perceptual nature of sound. Absolute Pitch May Not Be So Absolute Sensitivity and Selectivity of Neurons in Auditory Cortex to the Pitch, Timbre, and Location of Sounds 10.2 Describe the basic anatomy of the ear and the difference between the outer, middle, and inner ear. Cochlear Animation 2-Minute Neuroscience: The Cochlea 10.3 Examine the nature of hearing loss and what can be done to improve hearing in hearing-impaired individuals. Temporary Deafness Can Impair Multisensory Integration: A Study of Cochlear-Implant Users Hearing Loss in Older Adulthood: What It Is and How It Interacts With Cognitive Performance Neuro-Music Therapy for Recent-Onset Tinnitus Seeing Less Helps the Brain Hear More Tinnitus: Why Won’t My Ears Stop Ringing? Cochlear Implants Redefine What It Means to Be Deaf Helping Those With Hearing Loss Get in the Loop Scientific American Frontiers Cybersenses (cochlear implant 3:45–13:56) New Closed-Captioning Glasses Help Deaf Go Out to the Movies DNA Illustrations/Science Source 11The Auditory Brain and Sound Localization Alberto Ruggieri/Illustration Works/Getty Images LEARNING OBJECTIVES 11.1 Discuss the ascending neural pathways from the cochlea to the auditory cortex. 11.2 Identify the processes our auditory system uses to localize the sources of sounds in space. 11.3 Explain the concept of auditory scene analysis and how it is achieved by the auditory system. 11.4 Interpret how biosonar allows animals to use active hearing to negotiate their environment. 11.5 Discuss factors that go into designing a concert hall with good acoustics. INTRODUCTION Some of you may know people, perhaps family members, who constantly misplace their cell phones. The routine for finding a missing cell phone is a well-known script. You get someone to call the number of the missing phone. Once you hear it start ringing, it is easy to locate the lost phone, perhaps under a pile of magazines moved out of the way earlier in the day. In this particular situation, we take a simple fact of our auditory system for granted—that we can use sound to localize objects in space. But how do we determine from the sound of the hidden phone exactly where it is located? Few of us reflect on this ability when we dig the phone out from under those magazines. But the localization of objects in space via the auditory system is a very complex process and one that we discuss at length in this chapter (Figure 11.1). At a perceptual level, we use an important cue to home in on the spatial location of the sound, namely, its loudness. The louder the phone sounds, the closer you are to its source. Thus, if the missing phone sounds louder in the kitchen than it did in the living room, you are getting closer. If it is louder still in the pantry, perhaps you left it by the rice when you were cooking and talking on the phone. Interestingly, however, we can localize objects by sound even when we are standing still. In this case, our auditory system does an amazing calculation of differences in loudness and timing between the two ears. Yes, if the cell phone is to your left, the sound will be louder in your left ear ISLE EXERCISES 11.1 Sound Localization Experiment 11.2 Interaural Time Differences 11.3 Interaural Level Differences 11.4 Auditory Scene Analysis 11.5 Doppler Shift 11.6 Architectural Space and Echoes 11.7 Testing a Concert Hall 318 Sensation and Perception than your right ear, and we can use such cues to help us localize sound. However, this requires rather interesting neural machinery. Our auditory system has developed a mechanism that allows this rapid and precise calculation to be made quickly enough to detect the minute lag between when auditory information reaches one of our ears and the moment it reaches the other, as well as the tiny differ- ences in loudness. These involve incredible neuronal connections called the calyx of Held, named after the German scientist Hans Held, who first discovered these synapses at the end of the 19th century. The calyx of Held is a super-giant synapse that connects two neurons in the auditory primary cortex and has been found to be responsible for spatial localization across a number of mammalian species (Yang & Xu-Friedman, 2013). Because these synapses are so big, they allow the extremely rapid transfer of infor- mation from one cell in the network to the next. This extremely rapid transmission of information is necessary for the millisecond-precise timing that is necessary for using the auditory system to detect spatial differences. Thus, you can thank your calyxes of Held for your ability to track down that missing cell phone. As with vision, the auditory system sends information from the receptor cells to various locations in the brain for processing. As we discuss these brain regions and their responsibilities within the auditory system, think about how these systems are similar to and different from the systems within the visual system. BRAIN ANATOMY AND THE PATHWAY OF HEARING 11.1 Discuss the ascending neural pathways from the cochlea to the auditory cortex. Auditory Nerve Fibers Inner hair cells in the cochlea form synapses with auditory nerve fibers. These nerve fibers are bundled together in the eighth cranial nerve to be sent to the brain. We describe the properties of auditory nerve fibers in this section. Any particular audi- tory nerve fiber has a characteristic frequency to which it is most sensitive, consistent with place code theory, discussed in Chapter 10. If you examine Figure 11.2, you will see that the nerve fibers have characteristic frequencies to which they respond best, but they also respond to sets of frequencies both higher and lower than these characteristic frequencies. This is referred to as their tuning curve, which consists of the range of tones the cell responds to and its peak characteristic frequency. In Figure 11.2, the graph depicts sensitivity, the lowest sound level that a particular nerve fiber cell will respond to. The lower the curve goes, the more sensitive that nerve fiber is to a particular frequency. Thus, we can see that the auditory system continues to code for frequency in the auditory nerve, similar to the way the optic nerve codes for spatial location. Now we will describe the path auditory information takes to the auditory regions of the cortex. Follow carefully along with Figure 11.3 as you read this section. The FIGURE 11.1 Where Is the Cell Phone? © iS to ck ph ot o. co m /a le xy tr en er 319 Chapter 11: The Auditory Brain and Sound Localization FIGURE 11.2 Tuning Curve for Eighth Cranial Nerve Fibers These curves show the maximum sensitivity to frequency for different eighth cranial nerve fibers. 80 Fiber 1 0 80 Fiber 4 0 80 Fiber 2 0 80 Fiber 3 0 80 80Fiber 5 Fiber 6 0 0 0.1 1.0 Frequency (kHz)Frequency (kHz) 10.00.1 1.0 10.0 0.1 1.0 10.00.1 1.0 10.0 0.1 1.0 10.00.1 1.0 10.0 auditory nerve fibers form part of the eighth cranial nerve, which then makes its way to the brain. The auditory part of the eighth cranial nerve takes a complex journey to the auditory cortex. The auditory tract goes through the cochlear nucleus first. Attached to the cochlear nucleus is the trapezoid body, which is important in determining the direction of sound. The cochlear nucleus also contains subnuclei with specific functions, such as sensitivity to the onset and offset of tones at particular frequencies. Other cells in the cochlear nucleus also serve a lateral inhibition function. These cells sharpen our detection of a particular frequency of incoming sound by inhibiting the response to nearby frequencies, either higher or lower. The tuning curve is made steeper on the sides, if you will. From the cochlear nucleus and the trapezoid body, the sound signal goes to the superior olive in the brain stem. The superior olive receives input from both ears. This early crossing over of information from each side of the auditory system is critical for sound localization. From the superior olive, the next synapse in the ascending pathway of auditory information is the inferior colliculus (plural colliculi). The inferior collicu- lus then projects to the medial geniculate nucleus of the thalamus. The medial genicu- late nucleus then projects to the auditory cortex. The medial geniculate nucleus projects to the cortex but also receives input back from the cortex (He, 2003). Indeed, there are more returning (or efferent) connections from the cortex to the medial geniculate Cochlear nucleus: a structure in the brain stem that receives input from the inner hair cells Trapezoid body: a structure in the brain stem that plays a role in determining the direction of sounds Superior olive: a structure in the brain stem that receives input from the inner hair cells and from the cochlear nucleus Inferior colliculus: a structure in the midbrain that receives input from the superior olive Medial geniculate nucleus: a structure in the thalamus that receives auditory input from the inferior colliculus and sends output to the auditory cortex 320 Sensation and Perception Right hemisphere Auditory cortex Auditory cortex Cochlea Cochlea Main pathways Secondary pathways Type I auditory nerve fibers from inner hair cells Superior olivary complex Superior olivary complex Trapezoid body Trapezoid body Lateral lemniscus La te ra l le m ni sc us Medial geniculate body (in thalamus) Medial geniculate body Inferior colliculus Inferior colliculus Contralateral Ipsilateral Auditory cortex Superior olivary complex Medial geniculate body Inferior colliculus Cochlear nucleus Cochlear nucleus Auditory nerve Auditory nerve M id br ai n M id br ai n B ra in s te m B ra in s te m Left hemisphere C or te x C or te x Th al am us FIGURE 11.3 The Auditory Pathway Information goes from the cochleae of the ears up the ascending pathway to the auditory cortex in both the left and right hemispheres. nucleus than there are ascending (or afferent) connections from the medial geniculate nucleus to the cortex. The area in the cortex projected to by the medial geniculate nucleus is called the primary auditory cortex, located in the temporal lobe of the brain. Auditory nerve fibers from each ear go to each side of the temporal lobe, but there are more from the right ear that go to the left temporal lobe and more from the left ear that go to the right temporal lobe (see Figure 11.3). Because both ears project to both hemispheres, it is often difficult to assess hemispheric specialization in the brain. Auditory Cortex The auditory cortex is a large multifaceted area located in the temporal lobe. It is located under the lateral sulcus along the top of the temporal cortex. Figure 11.4 shows the complex organization of the auditory cortex. The first area that receives input from the medial geniculate nucleus is known as the primary auditory cortex (also known as A1 or area 41). Cells in the primary auditory cortex show a tonotopic organization. That is, cells show a maximal response to specific frequencies, and these cells are orga- nized in maplike patterns. The primary auditory cortex is one of three areas that make up the auditory core region. The other two are the rostral core and the rostrotemporal core. Another area within the auditory cortex surrounds the auditory core region. The belt and parabelt regions essentially wrap around the primary auditory cortex, hence the reference to belts, but are not part of the core region. All of these regions can be seen in Figure 11.4. Figure 11.4 also shows the tonotopic organization of the auditory cortex. From the basilar membrane on, the auditory system is coded in terms of the frequencies Auditory cortex: the areas in the temporal cortex that process auditory stimuli Primary auditory cortex: the first area in the auditory cortex, which receives input from the medial geniculate nucleus Tonotopic organization: the organization of neurons within a region in the brain according to the different frequencies to which they respond Auditory core region: an area of the auditory cortex, consisting of the primary auditory cortex, the rostral core, and the rostrotemporal core Rostral core: an area in the auditory core region of the auditory cortex Rostrotemporal core: an area, in addition to the rostral core, in the auditory core region of the auditory cortex 321 Chapter 11: The Auditory Brain and Sound Localization of sounds in the environment. Within the auditory core regions, one can find a tono- topic organization of the tissue within this region, gradually moving from high to low frequencies. Thus, just as the visual system is organized by space, we see frequency cod- ing throughout the auditory system. The core region appears to serve the same function as V1, allowing for the primary analysis of fre- quencies. The belt and parabelt regions seem analogous to the extrastriate cortex, doing more complex analyses of the auditory sig- nal. For example, neurons in these regions are less responsive to pure tones at particular frequencies than they are to complex stimuli consisting of multiple frequencies. “What” information and “where” infor- mation are separated in the auditory system in a manner analogous to the visual system (Figure 11.5). With respect to audition, the “what” system is involved in using audi- tory information to identify the identity of a sound. The “what” system forms the basis of both speech perception and music perception, which we consider in Chapters 12 and 13, respectively. The “what” system starts in the core region and then moves to more anterior parts of the tempo- ral lobe. In contrast, the “where” system is responsible for local- izing sound in space, a topic for this chapter. It is this system that allows us to determine that the alto saxophone is coming from one side of a jazz band, but the electric bass is coming from the other. The “where” system also begins in the core region of the auditory cortex and then moves to posterior regions of the tem- poral lobe as well as in the posterior parietal cortex (Rauschecker & Scott, 2009). TEST YOUR KNOWLEDGE 1. Describe the pathway as information leaves the cochlea and heads to the cortex. 2. What is meant by the concept of the “what” and “where” systems in the auditory cortex? How does each contribute to our understanding of the auditory cortex? LOCALIZING SOUND 11.2 Identify the processes our auditory system uses to localize the sources of sounds in space. Thousands of people have learned to scuba dive, and they flock to places like Florida and the Bahamas to swim underwater among tropical fish and corals. Scuba diving is a very A1 Rostral core Auditory cortexLateral sulcus Temporal lobe Rostrotemporal core Belt Parabelt Low High Transverse temporal gyrus (Heschl’s gyrus) Right hemisphere Left hemisphere Auditory core region Auditory cortex Characteristic frequency FIGURE 11.4 Auditory Cortex The auditory cortex is located under the lateral sulcus along the top of the temporal cortex. The auditory cortex is composed of the primary auditory cortex as well as the rostral core, the rostrotemporal core, and the belt and parabelt regions. Parietal lobe Occipital lobe Frontal lobe Temporal lobe Auditory core region “W hat ” “W he re ” FIGURE 11.5 The “What” and “Where” Systems in the Auditory Cortex The “what” system is responsible for identifying the sources of sounds, whereas the “where” system is responsible for determining where in space those sounds are coming from. The “what” system runs along the dorsal areas in the temporal lobe and then into the frontal lobe. The “where” system heads from the auditory cortex to the parietal lobe. Belt: a region of the auditory cortex that wraps around the auditory core regions Parabelt: a region of the auditory cortex, in addition to the belt area, that wraps around the auditory core regions 322 Sensation and Perception safe activity, if done correctly and at relatively shallow depths. Perhaps the biggest danger for scuba divers is swimming too close to the surface when motorboats are about. For this reason, scuba divers are taught to listen for the sounds of boats and to associate that sound with danger. The difficulty in hearing while completely immersed is that sound travels much faster in water than it does in air. This creates difficulty in localizing the direction of sound underwater. When a scuba diver hears a boat, it is much harder for him to deter- mine if the boat is coming from the left or right relative to his position. It is also more difficult to determine if the boat is in front of or behind the diver. This is all a direct function of the speed of sound in water, which throws off the mechanisms we use for sound localization, which are based on differences in timing and loudness between the left and right ears. Because underwater, those dif- ferences are less, sound is harder to localize in space. On land, however, we can make very clear judgments as to whether a sound is coming from the left or right. To test yourself on this dimension, go to ISLE 11.1. In this section, we look at the percep- tual and physiological mechanisms necessary for sound localization. In vision, spatial localization is direct and the main feature of cod- ing in the visual cortex. Spatial position is determined by the retinal position of an image. In the auditory system, the cochlea is organized tonotopically, and spatial local- ization is done by several indirect mechanisms. In the auditory system, sound local- ization is based on the comparison of sound in the two ears and is thus analogous to stereoscopic vision. To localize an object in space, we must know if it is to the left or right of us, whether it is in front of or behind us, and whether it is above or below us. That is, we must be able to localize sound in three-dimensional space. The azimuth refers to the left-right or side-to-side aspect of sound localization. Elevation refers to the up-down dimension of sound localization, and distance refers to how far a sound is from us, and whether it is in front of or behind us (Figure 11.6). We discuss how we determine each of these dimensions in this section. Interaural Time Difference The interaural time difference is the time interval between when a sound enters one ear and when it enters the other ear. In principle, this is a rather straightforward concept. A sound coming to us from the left will enter our left ear a split second before it enters our right ear. A sound coming to us from the right will enter our right ear a split second before it enters our left ear. Because our auditory system can detect this millisecond difference in timing, we can use the interaural time difference to determine if a sound is coming from the left or right. Thus, the interaural time difference gives us the location of the object along the azimuth. ISLE 11.1 Sound Localization Experiment Azimuth: the left-right or side-to-side aspect of sound localization Elevation: the up-down dimension of sound localization Distance: how far a sound is from the listener and whether it is in front of or behind the listener Interaural time difference: the time interval between when a sound enters one ear and when it enters the other ear FIGURE 11.6 Sound Localization The azimuth refers to the left-right or side-to-side position along the horizontal plane. Elevation refers to the up-down dimension along the median plane. Distance refers to how far a sound is from us and whether it is in front of or behind us. The horizontal plane bisects the head along the left- right dimension, whereas the median plane bisects the head along the vertical axis. Median Plane Distance Horizontal Plane Elevation Azimuth 323 Chapter 11: The Auditory Brain and Sound Localization To determine the actual interaural time difference is relatively straightforward, although it involves a bit of arithmetic. The speed of sound in air is 343 m/s, and the left ear and right ear are about 10 cm apart. If a sound is coming from a source directly to our right, the sound will reach the right ear 640 microseconds (0.00064 seconds) before it reaches the left ear. That’s an amazing 640-millionths-of-a- second difference our ears are detecting! However, not all sounds will be coming directly from the left or the right. Some may be slightly in front on one side or slightly behind (Figure 11.7). We can think of a big circle around the head and measure the angle of the sound to our head in degrees. If the sound is coming from 90 degrees, that is, directly to our left or right, the time difference will be 640 micro- seconds. However, if the angle is only 20 degrees (i.e., coming from in front of us but slightly to the right), the time difference will be only 200 micro- seconds. Thus, the longer the time lag is between the ears, the more toward 90 degrees (or immediately right or left) the sound is coming from. Figure 11.7 shows this point graphically. Research from psycho- physical experiments confirms that people do detect such incredibly small time lags (Fedderson, Sandel, Teas, & Jeffress, 1957). For a region in the brain to compute the interaural time difference, it must have access to information entering the auditory system from both ears. It is likely that for this reason, early crossing over of information from one side of the brain to the other evolved within the auditory system. Physiologically, interaural time differences appear to be computed in the medial superior olives. Yin and Chan (1990) found neurons within the medial superior olives of cats that were sensitive to differences in the inter- aural time difference, providing support for this region of the brain as being critical for sound localization (Figure 11.8). Go to ISLE 11.2, both parts, to experience and see how interaural time differences influence the direction from which you hear a sound. FIGURE 11.7 Interaural Time Differences At different angles relative to the head, a sound will reach the left and right ears with different interaural time differences. The auditory system can compute these time differences and use them to determine the direction of a sound. 90° ITD = 640 µs 60° ITD = 480 µs 120° ITD = 480 µs -60° ITD = 480 µs -20° ITD = 200 µs -160° ITD = 200 µs 160° ITD = 200 µs 20° ITD = 200 µs 180° ITD = 0 µs 0° ITD = 0 µs -120° ITD = 480 µs -90° ITD = 640 µs FIGURE 11.8 Interaural Time Difference in the Medial Superior Olives Information from the cochlea makes a synapse in the cochlear nucleus and then travels to the medial superior olives, where information from the right and left cochleae is first combined. Right cochlea Sound source Auditory nerve Brain stem Cochlear nucleus To higher brain centers From right cochlea From left cochlea LSOLSO MSO MNTB ISLE 11.2 Interaural Time Differences 324 Sensation and Perception Interaural Level Difference The interaural level difference is the difference in loudness and frequency distribution between the two ears. As sound travels, its strength dissipates. For example, if you are very close to a loud sound, it will sound loud to you. But if you are some dis- tance from the same loud sound, it will not be as loud. Think of a dog barking. If it is right in front of you, it will sound loud, but if the same dog barks across a big grassy field, the sound will be much less loud to you. Amazingly, our ears can detect loudness differences between the left and right ears. However, more important for sound localization is that the head casts an acoustic shadow, which changes the loudness and frequency distribution of sound going to each ear. We can define the acoustic shadow as the area on the side of the head opposite from the source of a sound in which the loudness of a sound is less because of blocked sound waves. The acoustic shadow is much more prominent for high-frequency sounds than it is for low-frequency sounds (Figure 11.9 and ISLE 11.3). Thus, sound coming from the left or right will have a different loudness in each ear and a different frequency composition in each ear. High-frequency sounds will be louder in the closer ear, although low-frequency sounds will be approximately the same loudness in each ear. That is, the ear that is closer to the sound will hear the sound as some- what louder, and that ear will also detect more of the higher frequency components of the sound. If it is a pure tone, then only the loudness difference will matter, but even with pure tones, the loudness dif- ference will be greater the higher the frequency of the sound. Psychophysical research shows that peo- ple are sensitive to these interaural level differences. Physiological studies suggest that interaural level dif- ferences are computed in the lateral superior olive. The Cone of Confusion The cone of confusion is a region of positions in space in which sounds create the same interaural time and interaural level differences. That is, imagine a sound that is coming from 45 degrees from below you and to the left. This sound will create the same differences as a sound that is coming from 135 degrees above you and to the right. If your head remains still, you cannot tell the difference in spatial position between sounds coming from these sources. This cone of confusion is depicted in Figure 11.10. Although the cone of confusion is real—it is easily demonstrable in psy- chophysical experiments—it is easily remedied. As soon as the head moves, the ears change position relative to the sources of sound, and sounds from different locations now may have different characteristics. So the cone of confusion seldom bothers a person in motion. FIGURE 11.9 Acoustic Shadow High-frequency sounds have shorter wavelengths than low-frequency sounds. This causes more reflecting or blocking of the high-frequency waves off the side of the head than occurs with low-frequency waves. Consider how big waves on the ocean roll right past objects such as buoys, but smaller waves break off when encountering such objects. It is the same principle here. When a sound wave encounters objects smaller than its wavelength, the sound wave bends around it, like the low-frequency sound does here. When a sound wave encounters an object bigger than its wavelength, the sound wave does not bend as much around the object. In fact, it may be blocked or distorted, as in the drawing. Low-frequency sound High-frequency sound Acoustic shadow ISLE 11.3 Interaural Level Differences Interaural level difference: the difference in loudness and frequency distribution between the two ears Acoustic shadow: the area on the side of the head opposite from the source of a sound in which the loudness of a sound is less because of blocked sound waves 325 Chapter 11: The Auditory Brain and Sound Localization Elevation Perception Sounds may be coming from above us and below us as well. For spatial localization, identifying this aspect of a sound is as critical as know- ing the angle of orientation of our head to that sound. Interestingly, the way our auditory sys- tem detects elevation is a function of changes in sound frequency created by the folds in our outer ears, the pinnae (Figure 11.11). The bumps and folds in the pinnae cause slight reverberations (i.e., echoes). These reverberations amplify some frequencies and dampen others. The spectral shape cue is the change in a sound’s frequency envelope created by the pinnae. It can be used to provide information about the elevation of a sound source. Experiments demonstrate that the shape of the pinna affects elevation perception. When the shape of the pinna is artificially altered, elevation perception is impaired. However, after sufficient time with a newly shaped pinna, ele- vation perception will be relearned and achieve normal levels (Hoffman, Van Riswick, & Van Opstal, 1998). Thus, one might expect some- one with an extreme ear piercing to temporar- ily lose some ability to localize sound, but that ability should return. Indeed, in Hoffman et al.’s (1998) study, participants with modified pinnae returned to normal localization after approximately 20 days (Figure 11.12). Detecting Distance Detecting the distance of a sound relies on a different set of principles. One of the chief methods for inferring the distance of sounds relies on our internal knowledge of the loudness of familiar sounds. Thus, when you hear a dog bark, you have knowledge of the approximate loudness of that specific dog’s barking, that breed of dog’s barking, or dog barking in general. Then, on the basis of the actual loudness of the bark, you can judge its distance as being close or far. Similarly, human voices can be judged as Axis of cone FIGURE 11.10 The Cone of Confusion Sounds coming from any location on the surface of the cone result in nearly identical interaural time and interaural level differences. Note that the cone is imaginary but exists relative to the sounds present, so there are many possible cones of confusion for different ranges of sounds at different distances and angles to the person. Cone of confusion: a region of positions in space in which sounds create the same interaural time and interaural level differences Spectral shape cue: the change in a sound’s frequency envelope created by the pinnae; it can be used to provide information about the elevation of a sound source FIGURE 11.11 The Human Pinnae Pinnae vary greatly from one person to the next. Each individual learns from birth how his or her pinnae change sound as it enters the ears. The shape of the pinna can provide information about the height of the source of a sound. © iS to ck ph ot o. co m /b ik ec © iS to ck ph ot o. co m /i ta lb y © C ha d Ba ke r/ Ja so n Re ed /R ya n M cV ay /P ho to di sc /T hi nk st oc k © C ha ts am /C C- BY -S A -3 .0 326 Sensation and Perception being closer or farther on the basis of their loudness. This, of course, may lead to errors. We may judge an unseen person’s voice to be near simply because he is speaking loudly. Frequency also plays a role in distance perception. High-frequency sounds show a greater decrease in loudness as a function of distance than do low-frequency sounds. Remember that you can hear the bass line of the music from the party across the lake, but not the voice of the singer. This is also apparent when we hear thunder during a thunderstorm. Thunder nearby has a distinctive “crack” that causes people to look up and cats and dogs to scurry for shelter. Distant thunder lacks the high-frequency component, and we hear it as a less-jarring “boom.” So, a diminishment of the high- frequency component of sound also tells us about distance. Another cue for distance is the proportion of direct sound to reflected sound. When the dog barks across the park from you, some of the sound you pick up travels directly from the dog to your ears. But some of the sound from the dog bounces off objects, such as the ground, the trees, and the wall by the tennis courts. This reverberant sound may also reach your ears. Close-by objects will have larger ratios of direct to reflected sound than faraway objects, under most circumstances. Reverberant sound will have slightly altered frequencies and interaural time differences than direct sound. This allows our auditory systems to compute the ratio, and this gives us information about the relative distances of sound sources. We return to the calculation of distance in our discussion of bat echolocation at the end of the chapter. For bats, detecting distance is easier because they can compare their own sound production with the return of echoes, allowing direct computation of distance. TEST YOUR KNOWLEDGE 1. What is the cone of confusion? Why is it not ordinarily a problem for listeners? 2. Why is the distance between the two ears important in localizing sound? How does having two ears contribute to our perception of localized sound? Might any of the cues to sound localization work in a person with only one functioning ear? AUDITORY SCENE ANALYSIS 11.3 Explain the concept of auditory scene analysis and how it is achieved by the auditory system. FIGURE 11.12 Ear Piercing Cu ltu ra C re at iv e (R F) /A la m y St oc k Ph ot o © iS to ck ph ot o. co m /a xe lb ue ck er t To m G ilk s/ A la m y St oc k Ph ot o 327 Chapter 11: The Auditory Brain and Sound Localization Stop for a moment right here and listen to all the sounds, however faint, that are cur- rently around you. What do you hear? Even in a quiet room, there are probably half a dozen different sounds you can distinguish. You might be able to hear the slight pulse of the air conditioner, the whirr of the refrigerator, and a car motoring past somewhere outside your window. You might hear the slight chatter from your sister’s iPod even though she’s wearing headphones, your mother on the telephone in the other room, and the distinct sound of your cat’s breathing while it sleeps nearby. Each of these sounds is made up of several frequencies, and in some cases, these frequencies may overlap (Figure 11.13). They are all hitting your cochlea simultaneously but seemingly not straining your ability to distinguish the sounds and localize them in space. This ability to distinguish the different sounds in the ambient environment is called auditory scene analysis (Bregman, 2005). There are only rare situations in which we cannot hear multiple sounds. Being in an environment with multiple sounds is the rule, not the exception. Thus, our audi- tory systems have evolved to separate and distinguish all these sounds despite the fact that the auditory stream from each sound may have overlapping frequencies as well as overlapping timing. Thus, the auditory system must group together those frequen- cies coming from common objects rather than simply grouping similar frequencies together. That is, even if the frequency of your whirring refrigerator is the same as that of your gently breathing cat, what is important is that we attribute each sound to its source. This is a distinctly different problem from the one we face with the visual system. When we look at objects, spatial information about them usually overlaps. Although Auditory scene analysis: the process of identifying specific sound-producing objects from a complex set of sounds from different objects at varying and overlapping frequencies FIGURE 11.13 Auditory Scene Analysis Complex sounds from five different objects enter a person’s cochlea at exactly the same time. The person’s ability to separate out these inputs and attribute them to the unique objects is known as auditory scene analysis. Refrigerator Frequencies produced by �ve different objects. All the sounds enter the ear simultaneously in a complex waveform. The auditory system breaks down the complex waveform and extracts individual sounds from it, known as auditory scene analysis. Music from an iPod Air conditioner Cat purring Person speaking 328 Sensation and Perception objects may be occluded, parts of objects tend to be near to one another. Because our visual system is organized in terms of space, scenes have a more inherent organization. However, given that our auditory system is organized tonotopically, the analysis of scenes is more complex, because any sound-emitting object is probably emitting sounds at multiple frequencies, some of which may overlap in terms of frequency with other sound-emitting objects. Bregman’s (1990, 2005) view of auditory scene analysis is very much akin to the principles of gestalt psychology. That is, the auditory system uses a number of heuris- tic rules to determine which frequencies go with which other frequencies and which sounds are associated with which objects. These rules are not perfect—they sometimes lead to errors in auditory scene analysis—but more often than not, they allow us to correctly parse the auditory landscape. Like the gestalt rules, these processes center on the ability to group different patterns of sounds together. Auditory scene analysis rules fall into three basic types. We group by timing—that is, sounds that are produced at the same time may be grouped together (temporal segregation). We group by space— that is, those sounds that are emanating from the same place in space are grouped together (spatial segregation). We also group by frequency—that is, sounds that are of the same fre- quency or harmonic pattern are grouped together (spectral seg- regation). We consider each in turn. You can also learn more about auditory scene analysis in ISLE 11.4. Temporal Segregation Consider the sound of your washing machine. There might be a relatively loud, high-fre- quency sound caused by the motor of your machine. At the same time, there might be a lower amplitude and lower fre- quency sound being produced by the rotation of clothes inside the machine. We group these sounds together because they are linked or correlated in time. When one stops, the other stops, and when the washing machine cycles again, both sounds start anew. In music, we group the violas together because they are all playing the same rhythms at the same time. This is the concept of temporal segregation. Sounds that are linked in time are grouped together, whereas sounds that are not cor- related with one another are not grouped together. However, in some situations, rapid alternation between two frequencies may be linked in time, but lead to the perception ISLE 11.4 Auditory Scene Analysis Short time (t) between occurrences of same tone Long time (T) between occurrences of same tone Perceived as a distinct auditory stream Perceived as a distinct auditory stream Time (a) B B B B A A A t t t t t t AF re q u e n cy Perceived as a single auditory stream Time (b) A A A T A B B B B F re q u e n cy TT TTT FIGURE 11.14 Temporal Segregation (a) In this sequence, because the high- and low-frequency tones alternate rapidly, we hear two separate sources of sound. (b) In this sequence, when the time between the successive tones gets longer, we join the two tones into a common theme and tend to hear one sound bouncing between two frequencies. Temporal segregation: the process whereby sounds that are linked in time are grouped together, whereas sounds that are not correlated with one another are not grouped together 329 Chapter 11: The Auditory Brain and Sound Localization of different streams rather than the same stream (Figure 11.14). Composers often use this rapid alternation of pitch in time to create the illusion of two sound streams from one musical instrument. Spatial Segregation In the section on sound localization, we discussed the mechanisms by which the auditory system can localize sound in space. To review, this involves comparing the loudness, timing, and frequency distribution of sounds reaching our left and right ears. Thus, with spatial segregation, we can distinguish where in space sounds are coming from. In auditory scene analysis, we can use this to group sounds together or separate them into different sources. If you can localize one frequency as coming from behind you and slightly to the right and another frequency as coming from above you and to the left, your auditory system will know that these frequencies are associated with different sound sources. However, if both a high frequency and a low frequency are coming from the same spot in space, you can assume that those frequencies are linked. Thus, when we hear the revving of the motor and the thump- ing of the stereo coming from the same source to our right, we associate both sounds with the Ford Mustang waiting to our right for the traffic light to turn. Moreover, when the light turns green, we hear all of these sounds getting softer together as the Mustang accelerates and drives away. Spectral Segregation Spectral segregation refers to a number of separate grouping rules concerning how we use frequency distributions to group sounds together. Note here that a source of sound, such as a person’s voice, has multiple frequencies present. You hear a person’s voice at its fundamental frequency, but that voice has a number of higher frequency harmonics. Indeed, any complex sound will be composed of multiple frequencies. Consider a person listening to another’s voice. The auditory system is detecting mul- tiple frequencies coming from the same location in space. Should the auditory sys- tem determine that the multiple frequencies are all coming from the same source, or should the auditory system determine that there are multiple objects in the same loca- tion, each emitting a different frequency? Obviously, in most cases, the former will be the case, not the latter. Similarly, consider a person listening to a musical perfor- mance. Each musical instrument has a fundamental frequency and a number of higher frequency harmonics. The listener must associate the fundamental frequency and the harmonics coming from the saxophone and segregate them from the frequencies com- ing from the piano in order to parse the musical phrases. Thus, a key aspect of spectral segregation is grouping by harmonic coherence (Figure 11.15). Harmonic coherence is a very strong predictor of what sounds our auditory system will group together. Consider listening to a recording of music. Through headphones or earbuds, you hear a large grouping of frequencies simultaneously. Even though there are only two sound sources—the speakers for your left and right ears—you hear a number of auditory streams, depending on what kind of music you are listening to. For example, if you are listening to a Mozart symphony, you will hear separate auditory streams for the violins, the cellos, the clarinets, the trumpets, and so on. If you are lis- tening to the Beatles, you will hear separate auditory streams for John, Paul, George, and Ringo. This grouping occurs even when some of the frequencies are coming from Spatial segregation: the process whereby sounds that are coming from the same location are grouped together, whereas sounds that are coming from different locations are not grouped together Spectral segregation: the process whereby sounds that overlap in harmonic structure are grouped together, whereas sounds that do not overlap in harmonic structure are not grouped together Harmonic coherence: when frequencies present in the environment resemble the possible pattern of a fundamental frequency and higher harmonics 330 Sensation and Perception 1,100 880 660 440 F re q u e n cy ( H z) Time 1st harmonic (fundamental) 2nd harmonic 4th harmonic 3rd harmonic Perceived as a single auditory stream (a) 5th harmonic 220 1,100 880 660 440 F re q u e n cy ( H z) Time 1st harmonic (fundamental) 2nd harmonic 4th harmonic nonharmonic tone Perceived as a distinct auditory stream Perceived as a single auditory stream (b) 5th harmonic 220 FIGURE 11.15 Spectral Segregation (a) When the fundamental frequency and higher harmonics of a particular sound, such as a voice, are presented together, the auditory system of the listener groups these tones together into a single stream from a single source. (b) When there is a tone present that is not possibly a higher harmonic of the fundamental frequency, that tone is perceived as a separate auditory stream against the background of the grouped frequencies. different sources. That is, harmonic coherence takes precedence over spatial segregation (Figure 11.16). TEST YOUR KNOWLEDGE 1. What is meant by the term auditory scene analysis? Why is auditory scene analysis important, and what processes help our auditory system carry it out? 2. What is harmonic coherence? What is the relation of harmonic coherence to spatial segregation? Auditory Development As most parents now know, the auditory system develops early. Indeed, studies show that shortly after birth, infants will respond differently to sounds they heard in utero and sounds they did not (Partanen et al., 2013). This does not mean that playing Mozart to your unborn child will make that child a genius, but it does suggest that the auditory system develops early. Findings show that the auditory system is functional at about the 25th week of gestation (i.e., the 25th week of pregnancy), long before the visual system is operational (Graven & Browne, 2008). Two-day-old infants can recognize the voices of their own mothers relative to the voices of other mothers (DeCasper & Fifer, 1980), which also sug- gests early development. However, hair cells and the auditory nerve will continue to change and develop throughout the first few years of a child’s life (Graven & Browne, 2008). In addition, experiments have shown that infants have equivalent thresholds across the frequency range as do young adults. They can detect soft sounds across the fre- quency range, at least from the age of 3 months, when it becomes possible to test thresholds. An important task for older children and adults is to separate concurrent sounds. Moreover, the research suggests that auditory localization of the sources of sounds occurs early. Young infants by the age of 3 months are already tuned to separate the sources of sounds in any auditory input (Grimault, Bacon, & Micheyl, 2002). In fact, this ability is often used to test visual abilities in young children. 331 Chapter 11: The Auditory Brain and Sound Localization Sound source A Fundamental frequency = 220 Hz F re q u e n cy ( H z) 4th harmonic 1,100 880 660 440 220 2nd harmonic 1st harmonic Time Fundamental frequency = 220 Hz F re q u e n cy ( H z) 5th harmonic 1,100 880 660 440 220 3rd harmonic + (a) + (b) Time F re q u e n cy ( H z) 5th harmonic 4th harmonic 1,100 880 660 440 220 3rd harmonic Perceived as a single auditory stream 2nd harmonic 1st harmonic Time Sound source B Sound source A Fundamental frequency = 220 Hz F re q u e n cy ( H z) 4th harmonic 1,100 880 660 440 220 2nd harmonic 1st harmonic (fundamental) Time Fundamental frequency = 200 Hz F re q u e n cy ( H z) 5th harmonic 1,100 880 660 440 220 3rd harmonic Time Sound source B Perceived as a distinct auditory stream Perceived as a distinct auditory stream FIGURE 11.16 Harmonic Coherence Takes Precedence Over Spatial Segregation (a) Harmonics of a common fundamental frequency are presented simultaneously but from different spatial sources. In this situation, the listener hears one auditory stream despite the spatial differences. (b) When the harmonic structure from two different sources is not coherent, we hear two separate auditory streams. EXPLORATION: Biosonar in Bats and Dolphins 11.4 Interpret how biosonar allows animals to use active hearing to negotiate their environment. You might not think bats and dolphins have much in com- mon. Bats are small, nocturnal flying mammals, whereas dolphins are large marine mammals (Figure 11.17). But they share one important aspect of their biology: Both use an active biosonar system. Biosonar is a process whereby animals emit sounds and then use comparisons of these emitted sounds and their returning echoes to sense the world around them. Biosonar has much in common with the electronic sonar systems human technology has devel- oped. The basis of both systems is the determination of the relation of emitted sounds and returning echoes (Simmons, 2012). This contrasts with our own hearing, which involves Biosonar: a process whereby animals emit sounds and then use comparisons of the emitted sounds and their returning echoes to sense the world around them 332 Sensation and Perception FIGURE 11.17 Bats and Dolphins Use Biosonar Bats and dolphins use their auditory system to navigate the world. But bats and dolphins have a biosonar sense. They emit loud, high-frequency sounds and then listen for the echoes. Both species, despite the difference in environments, use this information for hunting. Bats hunt small insects, and dolphins hunt fish. listening for external sounds but does not involve creating echoes with our own voices. For this reason, biosonar is thought of as a different perceptual system than our audi- tory system. But because it involves the basic processes of hearing, we include it here. For those who find this section intriguing, we recommend the book Sensory Exotica by Howard Hughes (1999), who discusses biosonar and a host of other nonhuman sensory systems. Bats hear sounds at much higher frequencies than we do. Most species of bats can hear sounds at over 100,000 Hz, whereas our highest frequency tops out at 20,000 Hz or below. Bats can produce very high-frequency sounds themselves. It is this combination that allows them to use biosonar. Because there is such a short wavelength (remember that frequency and wavelength are inverses), high-frequency sounds will be influenced by small objects that obstruct the wave pattern. For instance, a mosquito flying by would likely not impact a low-frequency wave, but it would deflect a high-frequency wave (Figure 11.18). That is, the low-frequency wave may miss the insect alto- gether, but a high-frequency wave will hit the insect and cause an echo to return to the bat. Thus, high-frequency sonar allows bats to detect small prey items, such as mos- quitoes, that lower frequencies would not pick up. The problem with high frequencies, though, is that they lose energy rapidly. This puts an additional requirement on animals using biosonar. Because high frequencies lose vol- ume rapidly, the initial sound must be very loud. Indeed, both bats and dolphins call very loudly. Bat calls are often as loud as 140 dB, and dolphin calls may reach 200 dB. Luckily, we humans are not disturbed by these super loud calls at night. Because these loud calls are made at frequen- cies we cannot hear, they can be going on all around us and we will not hear them. In essence, bats and dolphins must call very loudly to give their calls a little more range to strike potential targets. Biosonar systems work by integrating the processing of the call with the processing of the echo. Bats make calls, and then their super-sensitive auditory systems listen for the echoes that bounce off the objects in front of them and return to their ears (see Figure 11.20). This system provides the bat with information about the distance, size, and speed of the target it is tracking. This system allows bats to zoom in on prey and, in some cases, avoid obsta- cles and bigger predators. Figure 11.19 shows the two dif- ferent ways bats make these calls. In Figure 11.20, you © iS to ck ph ot o. co m /D on hy pe © iS to ck ph ot o. co m /G lo ba lP FIGURE 11.18 High-Frequency Waves Are Better for Capturing Information (a) The low-frequency wave passes right through the small object, which, for a bat, might be a tasty insect. The low-frequency wave will not detect the insect. (b) The high-frequency wave will be deflected by the small object, causing an echo to return toward the source of the sound. It is for this reason that animals equipped with biosonar use high-frequency sounds. 333 Chapter 11: The Auditory Brain and Sound Localization can see how a bat, once it detects an insect, uses its calls to home in on the insect. A bat needs to determine the distance from itself to an object. This is important—if the insect is too far away, the bat may not be able to catch up. If the object is close by, the bat may need to begin its capture behaviors. Distance from an object is determined by a calculation of the amount of time it takes for an echo to return to the bat. If the echo returns after a very short lag, the bat knows the object in front of it is very close. If the echo takes longer to return to the bat, the object is farther away. In bat (and military) circles, this aspect is known as the target range, that is, how far away the target is. This is a quite remarkable ability when you consider how close bats are likely to be to their targets and when you con- sider how fast sound moves. Given that sound travels at 343 m/s, it will take only 12 ms between a bat’s call and the echo’s return from a target 2 m away. Yet bat auditory systems can track such incredibly fast echoes. A target farther away than 2 m will take more than 12 ms, and a target closer will take less time. Thus, by timing the echo, the bat receives a precise estimate of the target range. Bats also need to know how large an object is. If a bat is following an owl, for example, it may wind up being the meal rather than finding a meal for itself. Thus, bats must be able to infer size information from their biosonar sys- tem. This information is also determined by the returning echoes. Larger objects produce bigger echoes. Thus, a bat can infer size from the relative loudness of the echoes. This determination is relative to the distance of the object, which is determined by timing. Thus, closer objects will produce bigger echoes too, but the bat’s auditory system can account for distance when using echo loudness to determine the size of the object. A bat needs to know if it is gaining on its target or falling behind in its pursuit. If the insect is flying faster than the bat, the bat may need to speed up. If the bat is flying faster than its prey, it may need to moderate its speed to be ready to engage in capturing behaviors. And if the bat is traveling at the same speed as the moth it is pursuing, it will not be able to catch it. This is normally not a problem, as bats fly much faster than most of the insects they prey on. But the bat still must know how fast it is approaching so that it can initiate its catching behaviors at just the right time. They determine their rate of approach by measuring something called the Doppler shift in their calls. Doppler shifts are apparent changes in frequency that occur when there is rela- tive motion between the source of a sound and a detector. In this case, both the source of the sound (the bat’s call) and the detector (the bat’s detection of echoes) come from the same place. The Doppler shift is an interesting phenomenon in and of itself, so we take a slight digression to explain it here. If a stationary object is emitting a sound at a particu- lar frequency, a stationary perceiver will hear it at that frequency. However, if the sound-emitting object starts FIGURE 11.19 Bat Biosonar Calls Constant frequency–frequency modulated (CF-FM) calls involve long calls at a single frequency. When an object is detected, the bat sweeps down in frequency. This is the characteristic pattern of forest bats. Frequency modulated (FM) calls involve just the sweep from high frequency down to relatively lower frequencies. This is the characteristic pattern of desert bats. CF-FM Call emiTemiT FM Call 90 60 30 3rd Harmonic 2nd Harmonic 1st HarmonicF re qu en cy (k H z) Components CF FM FM only B. Detect and pursue D. Resume search Return to search mode C. Capture No calls at moment of capture 0.05 second “Feeding buzz” A. Search Call in search mode FIGURE 11.20 How Bats Use Biosonar Bat biosonar is used for the pursuit and capture of insect prey. The bat uses its call and the returning echo to locate food sources. Target range: the distance of a predator from its potential target, determined by timing an echo’s return Rate of approach: the measure of whether a predator is approaching a target or receding from it 334 Sensation and Perception moving away from the perceiver, the object will still be emitting at the same frequency, but each sound wave will be coming from slightly farther away from the perceiver. This creates a perceived frequency somewhat lower than that of the actual sound. Similarly, if the sound-emitting object is moving closer to the perceiver, the sound waves will essentially bunch up together, creating the experience of a higher frequency sound to the perceiver. This is why an approaching train seems to get higher in pitch, and a receding train appears to get lower in pitch to the station- ary observer, waiting for the train to pass. A person on board the train will not hear this shift because she is mov- ing with the train and therefore hearing it at the frequency at which the train sounds are emitted. Now think about one of these bats calling as it flies though the sky. Its calls come at a constant frequency, say 90,000 Hz. First, consider the bat is moving faster than the object, but the bat is moving in the same direction as its calls. Thus, relative to an object in front of the bat, the frequency will appear slightly higher than the frequency the bat is actually calling at because each call comes slightly closer to the pre- vious call. If the object is moving faster than the bat is, it will take longer for the waves to catch up with that object, and the frequency will be perceived as lower than the actual call (Figure 11.21 and ISLE 11.5). In the case of the bat, its auditory system knows exactly what frequency it is emitting. If the returning echo appears to be slightly higher than the frequency of the call, the bat knows that it is gaining on its target. Thus, if it hears the echo at 90,050 Hz, the bat knows that it is approaching the target. If the returning echo appears to be slightly lower in frequency than the frequency of the call, then the bat knows it is losing ground on its target. Thus, if hears the echo at 89,950 Hz, the bat knows that it is falling back from its target. Finally, a bat needs to know if its target is moving to the left or right and up or down relative to the bat’s position. Left-right determination is done by a comparison of the returning echo to the two ears, similar to the discussion of sound localization earlier in this chapter. If the echo is of equal amplitude in each ear, the bat knows the tar- get is straight ahead. However, a difference in amplitude between the left and right ears indicates that the target is to the left or to the right. If the object is to the left of the bat, the echo will be slightly stronger in the left ear than the right ear, and if the object is to the right, the echo will be slightly stronger in the right ear than the left ear. A bat also needs to know if its target is rising or falling relative to its flight pattern, that is, whether the target is flying higher or lower than the bat. This turns out to be the most complex computation by the bat’s auditory system, and it involves comparing sound inputs distributed across its complexly shaped pinnae. Bats have large movable pin- nae, so they are much better at elevation determinations than humans are. Studies have shown that blocking a bat’s pinnae interferes with its determination of the elevation of its target. Using this information, a bat uses its auditory system to “see” objects in the external environment (Simmons, 2012). Because it is using its auditory system, a bat can use this system to guide its flight in complete darkness. Biosonar also allows bats to be highly social. It has been shown that they use their sonar systems to avoid bumping into one another in the tight confines of caves, and some bats also use sonar to identify their own offspring (Bohn, Smarsh, & Smotherman, 2013). What about dolphins? Dolphin sonar is also a call-and- echo system (Hughes, 1999). Dolphins also use loud, high-frequency pulses. Indeed, dolphin calls have registered at over 200 dB and nearly 200,000 Hz! Dolphins com- pute distance to target, approach to target, size of target, and left-right and up-down differences in the same way as bats. What makes the dolphin’s system remarkable is that sound travels 4 times faster through water than it does through air. Thus, the dolphin’s brain must be even more sensitive to minute millisecond time differences than that of the bat. LL 1 1 • • • 5 2 3 4 5 λ Listener Sound Source FIGURE 11.21 Doppler Shifts If a sound source is approaching you, it will sound slightly higher in frequency to you than it would to an object moving along with it. Similarly, if a sound source is receding from you, it will sound slightly lower in frequency to you than it would to an object moving along with it. Because bats know the frequencies at which they are calling, they can use these Doppler shifts in the echoes to calculate rate of approach. ISLE 11.5 Doppler Shift 335 Chapter 11: The Auditory Brain and Sound Localization APPLICATION: Concert Hall Acoustics and Hearing 11.5 Discuss factors that go into designing a concert hall with good acoustics. Many people enjoy live music. People gather in large num- bers to hear concerts of varying types. In many forms of popular music, a concert may be played in a large stadium. Such stadiums are usually not designed for live music, and typically music is amplified through a variety of electronic means. This means that the sound that the audience hears is not necessarily coming from the direction of the stage. It is coming from wherever there are speakers projecting that amplified sound. In very large stadiums, such as the half-time show at a football game, such amplification is necessary because no single human voice could carry across such a stadium. Moreover, with the relatively slow speed of sound, a fan in the upper part of the stadium might detect a difference between the visual movements of the singer’s mouth and the music that is reaching him. However, in “classical” orchestral music, the music is not usually elec- tronically amplified. In these concerts, the audience hears the music emanating from the instruments on the stage, at the sound level of the actual playing of the instruments or the voices of the choir. For this reason, the design of the concert hall is important (Figure 11.22). The concert hall must be designed to deliver that music to the audience in a conducive way without unnecessary echoing or distorting the sound. This is especially important in the larger concert halls, where sound must reach people in the back at suffi- cient loudness. Architectural acoustics is the study of how physical spaces such as concert halls affect how sounds are reflected in a room. Listen to ISLE 11.6 for an illustration of how the physical space alters a sound. A concert hall wants the sound to come directly to audience members without much echoing or reverberation. Direct sound has advantages. For example, because lower pitches have lower frequency, they will tend to move around objects, such as chairs, that can cause echoing, more so than higher pitches. Thus, direct sound carries more of the intended musical pitches than does indirect sound. Because it is impossible to eliminate all indirect sound, it is also important to design the concert hall such that the indirect sound is as muffled as possible. Typically, architects try to create a 60-decibel drop from the direct sound to the first reverberation of indirect sound (Lokki & Patynen, 2015; Sabine, 1922). Indeed, Sabine introduced the concept of reverberation time to represent the difference between the onset of direct sound and the first onset of indirect sound. In some cases, it is defined differ- ently, but we go with Sabine’s original definition. Therefore, room acoustics tells us how sound travels from instruments to the listeners in the audience. The auditory system of the listener depends on the direction of the sound and the fre- quencies of the sound. Because higher frequencies are more influenced by room dynamics, architects must design spaces to especially reduce the reverberation of those higher fre- quencies (Patynen & Lokki, 2016). It is also important to design a concert hall in which the reverberation time is least intrusive. Most concert halls look for a reverber- ation time of about 2 sec- onds because this allows the music to be heard as intended and not distorted ISLE 11.6 Architectural Space and Echoes Architectural acoustics: the study of how physical spaces such as concert halls affect how sounds are reflected in a room Reverberation time: the difference between the onset of direct sound and the first onset of indirect sound FIGURE 11.22 Concert Hall People gather in such halls to hear musical performances. Acoustical engineers design these halls to channel the sound in order to make the music enjoyable and without unnecessary echoing or distorting the sound. Shutterstock.com /m ichaeljung 336 Sensation and Perception or essentially blurred by the presence of echoes. Recording studios will often use an even shorter reverberation time to allow for very precise recording (Beranek, 2003). An important factor in architectural acoustics is the expected loudness of the concert hall. As the loudness of the music increases, the importance of good acous- tics increases. At softer volumes, there is less reverbera- tion off the walls and ceiling of the concert hall, but this increases as the loudness increases (Meyer, 2009). Indeed, Meyer suggests that one way to test a concert hall is to examine the pattern of reverberation when loud music is presented. Most concert halls allow for softer sounds to sound pure, but as the music gets louder, the acous- tics become more critical. For this reason, it is import- ant to test each concert hall while it is under construction to ensure a high-quality sound (Beranek, 2003). ISLE 11.7 illustrates this process for an actual renovated concert hall. The next time you have the pleasure of hearing music live in a concert hall, look at the walls and the ceiling. They are designed to muffle the indirect sound as much as pos- sible. Absorbing materials such as fiber boards make up the odd-looking material on the walls and ceiling. Take note of the loudness level before the concert starts. Even though there might be more than 1,000 people there, you can prob- ably talk with your friends or family in a normal voice and be heard. Contrast that with a busy restaurant with typical walls without soundproofing. Even with only a dozen or so people in a restaurant without soundproofing, the loudness level may interfere with the ability to hear each other. You might think that some restaurants would add soundproofing so that their customers could converse better. But restaurants find that at noisier levels, customers order more and leave earlier, so unlike a concert hall, in which good acoustics is critical, restaurants might actually prefer the extra sound. ISLE 11.7 Testing a Concert Hall CHAPTER SUMMARY 11.1 Discuss the ascending neural pathways from the cochlea to the auditory cortex. Hair cells send signals into the auditory nerve, which then leave the ear and head toward the brain. The cochlear nucleus is a structure in the brain stem that receives input from the inner hair cells. The signal is then sent to the superior olive and from there to the inferior colliculus. The inferior colliculus sends the auditory signal to the medial geniculate nucleus, which then sends its output to the auditory cortex. The primary auditory cortex (and rostral core) is the first area in the auditory cortex that receives input from the medial geniculate nucleus. It is surrounded by the belt and parabelt regions. Tonotopic organization means that the auditory cortex is organized by frequency. The auditory cortex develops early in the fetus’s neural development. 11.2 Identify the processes our auditory system uses to localize the sources of sounds in space. To localize an object in space, we must know if it is to the left or right of us, whether it is in front of or behind us, and whether it is above or below us. Sound localiza- tion is a very important aspect of the auditory system, as it is often important to pinpoint a sound’s source. That is, we must be able to localize sound in three-dimensional space. The interaural time difference is the time interval between when a sound enters one ear and when it enters the other ear. The interaural level difference is the differ- ence in loudness and frequency distribution between the two ears. The acoustic shadow is the area on the side of the head opposite from the source of a sound in which the loudness of a sound is less because of blocked sound waves. The cone of confusion is a region of positions in space in which sounds create the same interaural time and interaural level differences. The shape of the pinna affects elevation perception, which has been demon- strated in experiments. When the shape of the pinna is artificially altered, elevation perception is impaired. This is known as the spectral shape cue. 11.3 Explain the concept of auditory scene analysis and how it is achieved by the auditory system. Auditory scene analysis is the process of identifying specific sound-producing objects from a complex set of sounds from different objects at varying and overlapping frequencies. The rules of auditory scene analysis fall into three basic types: temporal segregation, spatial segrega- tion, and spectral segregation. The auditory system devel- ops early. Indeed, studies show that shortly after birth, Chapter 11: The Auditory Brain and Sound Localization 337 infants will respond differently to sounds they heard in utero and sounds they did not. 11.4 Interpret how biosonar allows animals to use active hearing to negotiate their environment. Bats and dolphins use a complex system known as biosonar to negotiate their environments. Biosonar is a system of active hearing. Animals produce high-fre- quency sounds at very loud volumes and then listen for the echoes of those sounds. By comparing the sound produced to the echo received, these animals can extract useful information such as the distance to an object and whether they are gaining on or losing that object. Both bats and dolphins use biosonar to avoid obstacles and hunt prey. 11.5 Discuss factors that go into designing a concert hall with good acoustics. A concert hall is a place where people gather to listen to music. For this reason, it is important to design concert halls with the best acoustics to allow people to enjoy the music. Architects design concert halls to reduce indirect sound, that is, the sound bouncing off walls and ceilings rather than coming directly from the musicians. The more such indirect sound can be reduced, the better the con- cert will be. REVIEW QUESTIONS 1. List all of the brain regions that make up the ascend- ing auditory pathway. Describe at least one dif- ference of this auditory pathway from the visual pathways to the brain. 2. What are the “what” and “where” systems in audi- tory perception? What brain regions are involved in each system? 3. What is meant by the term tonotopic organization of the auditory cortex? 4. How can the interaural time difference be used to compute the spatial coordinates of a sound in the external world? 5. Why are interaural level differences more useful in localizing high-frequency sounds than low-fre- quency sounds? 6. What is the cone of confusion? How does it affect sound localization for an unmoving perceiver? 7. What are the mechanisms human auditory systems use to detect the distance of a sound? Why are we better at judging distance for a familiar sound? Why are bats and dolphins better at detecting distance? 8. What is meant by the term auditory scene analysis? What mechanisms do we use to group frequencies together from common sources? 9. What is the difference between a call and an echo in bat echolocation? Why does a bat need both to be able to detect objects? 10. How does a bat compute its rate of approach to an object? How might a bat be fooled by an insect equipped with the ability to produce high-frequency sounds but not the ability to fly faster than the bat? PONDER FURTHER 1. Can you relate specific gestalt principles to spe- cific ideas of auditory scene analysis? For example, how would the “law of simplicity” map onto spatial segregation? 2. Consider a difference in the visual and auditory sys- tem. The crossing over from the left and right sides of the world takes place much earlier and much more often for the neural processing of audition than vision. What advantages does this crossover provide for audition? Sensation and Perception338 KEY TERMS Acoustic shadow, 324 Architectural acoustics, 335 Auditory core region, 320 Auditory cortex, 320 Auditory scene analysis, 327 Azimuth, 322 Belt, 320 Biosonar, 331 Cochlear nucleus, 319 Cone of confusion, 324 Distance, 322 Elevation, 322 Harmonic coherence, 329 Inferior colliculus, 319 Interaural level difference, 324 Interaural time difference, 322 Medial geniculate nucleus, 319 Parabelt, 320 Primary auditory cortex, 320 Rate of approach, 333 Reverberation time, 335 Rostral core, 320 Rostrotemporal core, 320 Spatial segregation, 329 Spectral segregation, 329 Spectral shape cue, 325 Superior olive, 319 Target range, 333 Temporal segregation, 328 Tonotopic organization, 320 Trapezoid body, 319 Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review, study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning. Learning Objectives Digital Resources 11.1 Discuss the ascending neural pathways from the cochlea to the auditory cortex. 11.2 Identify the processes our auditory system uses to localize the sources of sounds in space. Auditory Spatial Processing in the Human Cortex Auditory Azimuthal Localization Performance in Water as a Function of Prior Exposure 11.3 Explain the concept of auditory scene analysis and how it is achieved by the auditory system. Audio Demonstrations Of Auditory Scene Analysis Tuning In to the Brain’s “Cocktail Party Effect” 11.4 Interpret how biosonar allows animals to use active hearing to negotiate their environment. Daniel Kish: How I Use Sonar To Navigate The World 11.5 Discuss factors that go into designing a concert hall with good acoustics. 10+ Tips for Designing Classrooms, Hospitals and Offices That Are Kind on Ears, From Julian Treasure One Man’s Quest to Find the “Sonic Wonders of the World” (start at 6:40) Alberto Ruggieri/Illustration Works/Getty Images 12Speech Perception Gusto/Science Source LEARNING OBJECTIVES 12.1 Discuss the complex process of speech perception and speech production with regard to the speech stimulus. 12.2 Sketch the mechanisms our speech perception system uses to extract coherent speech from the speech signal. 12.3 Contrast the different types of theories of speech perception, noting their similarities and differences. 12.4 Diagram the areas of the brain involved in speech perception and what happens when they are damaged. INTRODUCTION There is no function of our auditory system more vital to human beings than speech perception. Speech perception is critical in interacting and socializing with the people around us. Unless you spend most of your time in your study writing textbooks, you are likely to spend most of your waking hours talking to and listening to other people. Other people constantly surround us, and they communicate with us using speech. You hear “What’s your name?” and this sound informs you that the person speaking is asking to interact with you. You hear somebody say, “What’s the best restaurant near here?” and you know that she wants your opinion. Speech is everywhere and vital for our everyday life (Figure 12.1). Moreover, from the auditory signal of a person’s voice, we extract not just the information the speaker is conveying but also information about the speaker, including gender, age, mood, native language, national background, regional background, educational level, and so on. And all of this information comes from the auditory speech signal. We need to be able to understand speech rapidly. The average speaker may talk at a rate of roughly four words per second. That is one word every 250 ms. This means our auditory system must convert the sound signal into meaningful language information at very rapid rates to allow us to gather all the information that we need from speech. Our auditory system uses every bit of available information to do this, including using cross- modal information from vision to help us understand speech. It also turns out that the speech signal is very complex, which further taxes our speech perception system. This chapter examines some of the processes that allow our auditory systems to quickly interpret very complex speech signals. ISLE EXERCISES 12.1 Coarticulation 12.2 Voicing-Onset Time 12.3 McGurk Effect 12.4 Familiar vs. Unfamiliar Languages 12.5 Phonemic Restoration Effect 12.6 Genie 12.7 Broca’s and Wernicke’s Aphasia 342 Sensation and Perception A situation many people find themselves in from time to time is a crowded room filled with many conversations taking place at once. This situation might occur in a busy restaurant, while watching a sporting event, while at a fra- ternity party, or at an old-fashioned cocktail party. Indeed, in the technical literature, this phenomenon is known as the “cocktail party effect.” In the cocktail party effect, you are listening to the speech of one person, but your attention is then distracted by the mention of your name by someone across the room. You may not have even been aware of this conversation, but the mention of your name can drag your attention from the conversation you are a part of to one across the room (Cherry, 1953). To researchers in the 1950s and 1960s, this observation suggested that our atten- tional mechanism does not screen out all perceptual input even when we are focused on one conversation. We covered attention in an earlier chapter, but we mention it here to note that our speech perception systems are capable of monitor- ing multiple speech inputs at the same time. This requires enormously complex processing in our auditory system to separate out the different voices coming from different directions. From the point of view of your auditory system, all of the voices you hear in different conversations are causing movements along your basilar membrane at the same time. Think of this: Every sound has one entry point into our auditory system, via the stapes pressing on the oval window. Thus, the auditory system must first be able to separate out the various voices as well as all other sounds that are all occurring at the same time. This problem of segmenting the auditory signal into its component parts is called auditory scene analysis (discussed in Chapter 11), and it is necessary to determine which particular sounds are coming from which source (Bregman, 1990). Even if no person is speaking to you as you read these words, think of the different sounds you can pick up. You can sep- arate out the sound of the radio playing music in the other room easily from the barking of the dog across the street. Similarly, even when multiple voices are talking at the same time, we can separate out which voice goes with which person and place those voices appropriately in terms of their locations. Recent research shows that an important com- ponent of scene analysis for speech is the familiarity of a voice. A familiar voice, such as that of a spouse, a child, or a parent, can be used to mark a particular signal to either be attended to or not. We are better at both with familiar voices relative to unfamiliar voices when listening amid a chatter of conversations (Johnsrude et al., 2013). Like many perceptual processes, speech perception seems effortless and fluent to adults, in this case, listening to fluent speakers of their native language. However, as with many perceptual processes, what our auditory system must do to achieve this fluent perception is remarkably complex. Indeed, computer speech-processing systems are only now becoming truly workable, and even so, they require enormous computing power. THE HUMAN VOICE AS STIMULUS 12.1 Discuss the complex process of speech perception and speech production with regard to the speech stimulus. Perceptual processes begin with a stimulus in the environment. In the case of human speech perception, that stimulus is a voice. Other people produce speech, which our © iS to ck ph ot o. co m /J oh nn yG re ig FIGURE 12.1 Listening to Language Is Critical People must listen to and understand others’ speech in most aspects of their lives. 343 Chapter 12: Speech Perception perceptual apparatus must hear and understand. Therefore, it is necessary to understand just a bit about the acoustics of human voices in order to begin our discussion of speech perception. We are sure you will not be surprised to learn that the human speech organs are quite complex, and many different parts of our anatomy play a role in producing the myriad speech sounds we make. Speech sounds are produced by movements of the vocal apparatus (Figure 12.2). Speech sounds are acoustic signals, which start as air pressure changes in the lungs that are then mod- ified by a number of structures in the neck and mouth that create the sounds we use as speech. First, air that produces the sound passes through the trachea (or windpipe) and into the larynx (also known as the voice box). The larynx is an important part in the vocal tract but not the only one necessary for speech. Vocal folds within the larynx open and close to change the pitch of the sound. However, parts of the vocal tract higher up from the trachea are necessary for the array of consonant sounds our voices must make. From the larynx, the air passes through the pharynx into the mouth and nose, where teeth, tongue, lips, and the uvula can affect the sound signal. The pharynx is the top portion of the throat. The uvula is tissue in the back of the mouth attached to the soft palate. Closing the uvula prevents sounds from going up through the nasal cavity, affecting the quality of the sound. The sound signal then exits the vocal apparatus through the mouth and the nose. These articulators are depicted in Figure 12.2. Vowels and Consonants Human vocal tracts have to be able to make two categories of speech sounds: vowels and consonants. These terms are equivalent to the categories one learns when learning to read. Vowels are represented in English by the letters a, e, i, o, u, and sometimes y. Consonants are the letters that surround these vowel sounds, such as the sounds rep- resented by the letters b, v, and p. Vowels and consonants have different origins within the vocal tract. Vowels are produced by unrestricted airflow through the pharynx and mouth. Changing the shape of the mouth creates different vowel sounds. Restricting the airflow in one place or another along the way up from the larynx produces consonants. For example, movements of the tongue restrict the sound and change the nature of a consonant. Perceptually, as well, vowels and consonants are two distinct categories. Vowels are produced by vibrations of the vocal cords, and specific vowel sounds (“oo,” “ah,” “ee,” etc.) are made by changes in the position of the oral cavity, particu- larly the shape of the mouth. If you speak aloud the standard vowel sounds in English, you will feel the shape of your mouth move. As we change the shape of our mouth and vocal tract, we change the resonances in the sound. That is, each vowel sound has a characteristic pattern of harmonics. The harmonics that are of highest amplitude are known as formants. Formants are the frequency bands with higher amplitudes among the harmonics of a vowel sound. Each individual vowel sound has a specific pattern of formants. Figure 12.3 shows the shapes of the mouth and the frequency spectra for two different vowel sounds. Note that the fundamental frequency is the same for each Hard palate Soft palate (velum) Uvula Pharynx Larynx Air from lungs Open Closed Alveolar ridge Teeth Nasal cavity Lips Oral cavity Vocal folds Vocal folds Tongue FIGURE 12.2 The Vocal Tract Human beings have a sophisticated anatomical system for producing language sounds. Speech is initiated with air being pushed out of the lungs into the larynx. Vocal folds in the larynx vibrate at different frequencies, producing the different pitches human voices are capable of. Changes in the position of the pharynx, oral cavity, tongue, lips, and teeth can then shape that sound into the particular speech sound the person is saying. Trachea (windpipe): the tube bringing air to and from the mouth Larynx (voice box): a structure that affects the pitches of sounds being made by the human vocal tract Pharynx: the top portion of the throat Uvula: a flap of tissue at the top of the throat that can close off the nasal cavity Vowels: speech sounds made with unrestricted airflow Consonants: speech sounds made with restricted airflow Formants: frequency bands with higher amplitudes among the harmonics of a vowel sound; each individual vowel sound has a specific pattern of formants 344 Sensation and Perception vowel. Vowels are distinguished from one another by their formant frequencies. Consonants are produced by restricting or closing the flow of air along the vocal tract. In English, the letters b, d, and f all rep- resent consonant sounds. Some sounds, such as “b” and “v” and “d” and “f,” are very similar and are therefore sometimes difficult to distinguish. Other consonant sounds used in English include “m” as in mom and “k” as in kick. In English, we use the consonant sound “sh,” though other languages, such as Spanish, do not. In contrast, Spanish uses sounds, such as the “ñ” in niño, which we do not typically use in English. Three physical features are important in determining the sound of a consonant. The place of articulation refers to the point along the vocal tract at which the airflow is constricted. The place of articulation may include the tongue, the lips, the teeth, and structures at the back of the mouth, such as the soft or hard palate. Try saying “b” versus “d.” The manner of articula- tion refers to how that restriction occurs. This involves whether the lips are pushed together, whether the tongue is at the front or back of the mouth, and other variants of the mouth’s position. Try saying a few simple words (e.g., “Bring the plate of choc- olates, but do not drop it on the floor”) or simple nonsense syllables (e.g., “b,” “f,” and “m”), and attend to the positions of your tongue, teeth, and lips as you say each sound, and you will see how each sound is accompanied by different positioning of the vocal tract. Finally, voicing refers to whether the vocal cords are vibrating or not. Think of the difference between the way we say the sound in the letter p and the sound in the letter b. When you say the letter b, you can feel your vocal cords vibrating immediately as you produce the sound. “P” is not accompanied by vocal cord vibration until after the “p” has been pronounced. We say that “b” is a voiced consonant and that “p” is an unvoiced consonant. “S” and “z” are also complementary unvoiced and voiced sounds. Figure 12.4 shows what our vocal tract is doing when we speak various sounds. Speech The purpose of speech is to convey information. When you tell your friend “I am hun- gry,” you are telling that person how you are feeling and likely implying what you want to do—have lunch. In order for your friend to understand what you are saying and your intentions behind it, he must be able to correctly perceive your utterance. Thus, speech perception involves the process of how we go from the sound signal coming from a speaker to understanding in a listener. This involves complex motor tasks in the speaker, as we just discussed, and complex auditory and cognitive tasks in the listener, which we turn to now. We start with phonemes. /i/ as in “bead” /æ/ as in “bad” Oral cavity (a) (b) Tongue Pharynx Oral cavity Tongue Pharynx F 2 F 2 F 1 F 1 F 3 F 3 Frequency (Hz) A m pl itu de (d b S P L) 0 0 5,000 20 40 60 Frequency (Hz) A m pl itu de (d b S P L) 0 0 5,000 20 40 60 FIGURE 12.3 Vowel Sound Production (a) The mouth producing two different sounds. (b) The frequency spectra for those sounds. Place of articulation: the point along the vocal tract at which the airflow is constricted Manner of articulation: how airflow restriction in the vocal tract occurs Voicing: whether the vocal cords are vibrating or not Voiced consonant: a consonant that is produced using the vocal chords Unvoiced consonant: a consonant that is produced without using the vocal chords 345 Chapter 12: Speech Perception Phonemes are the basic units of sound in human language. When a phoneme is changed, the meaning of the utterance is changed. Consider the word mop, which consists of three pho- nemes: “m,” “ɒ” (the international symbol for the vowel sound in mop), and “p.” If we change the first phoneme from “m” to “h,” we get a new word, hop, with a different meaning. If we change the vowel sound in the middle from “ɒ” to the “æ” sound in map, we also get a different word. Finally, we can change the “p” sound at the end of the word to a “b” sound and get mob. Thus, in this case, each of the three letters in the word mop represents a separate phoneme. It is important to note that phonemes refer to sounds, not letters. We must represent some phonemes with letters in this book, but any particular letter may stand in for multiple phonemes. Think of the letter combination ch, which may represent the first sound in chil- dren and the first sound in choir, despite the fact that the two sounds are quite different and it is two letters being used for one phoneme in each case. Figure 12.5 shows a classic diagram of how we use the mouth to make many different phonemes. In English, as in many languages, convention in spell- ing does not match phonemes in a straightforward and direct manner. Users of English must learn by rote what vowel sound to use when pronouncing such words as mood and blood, rough and through, and heard and beard. In other cases, we may have two separate symbols that represent the same phoneme, such as in the “f” sound “pit,” “bit” /p/, /b/ “fine,” “vine” /f/, /v/ “tip,” “dip” /t/, /d/ “kit,” “get” /k/, /g/ FIGURE 12.4 Articulation of Consonants Consonant sounds depend on the place of articulation, manner of articulation, and voicing. Several consonant sounds are depicted here. FIGURE 12.5 Production of Vocal Sounds Seventeenth-century diagrams of the anatomy of the human throat, showing how vocal sounds are produced. Each of the smaller 34 diagrams shows the vocalization of a particular letter or sound. The larger diagram (lower right) is labeled with the epiglottis, the larynx, the trachea (aspera arteria), and the esophagus. These diagrams are from An Essay towards a Real Character and a Philosophical Language (1668) by English clergyman and natural philosopher John Wilkins (1614–1672). In this book, Wilkins attempted to establish a universal language. Phonemes: the basic units of sound in human language 346 Sensation and Perception TABLE 12.1 Phonemes of American English Using the International Phonetic Alphabet Vowel International Phonetic Alphabet Example Words a ace, stay a hat, cat a art, father ay/ei play, eight e pet, red e he, key ear/ere fear, here ere/air there, pair i sit, mit i pipe, life ow, ou how, pout o go, home o spot, lock ou/a thought, ball tour, pure oi/oy join, toy oul/u could, put oo/ue food, blue our/ure ur/ear burn, earn Consonant International Phonetic Alphabet Example Words b bat, jab c cab, block ch choke, church d do, bad f fan, if g game, log h hot, help j jump, large l let, liftel m map, lime n note, pen n sting, fi nger p pool, tap r rest, train s soap, kiss s treasure, vision sh shirt, trash t tip, setting th thought, bath th this, brother v vote, hive w win, willow y yell, yet z zoom, crazy International Phonetic Alphabet: an alphabetic convention that provides a unique symbol for each and every phoneme in use in human languages in father and phoneme. Thus, spelling and written letters only partially capture the richness of phonemes. Linguists, in studying phonemes, require a system in which a single symbol rep- resents each sound that people make in a particular language. This alphabet is known as the International Phonetic Alphabet, which provides a unique symbol for each and every phoneme in use in human languages. The International Phonetic Alphabet includes symbols for the hundreds of unique sounds that are used across the thousands of languages in use today (MacMahon, 1996). Table 12.1 shows the phonemes used in English and the symbols used to represent them in the International Phonetic Alphabet. Inspect Table 12.1 carefully. You will see that first, in contrast to what you were taught in elementary school, the English language has 15 vowel sounds and 24 consonant sounds in what is considered standard American English, though there is a lively debate on whether there is a standard American English. Because we use only 26 symbols in 347 Chapter 12: Speech Perception English to represent 39 sounds, many individual letters and letter combinations (e.g., sh, oo) must be used to represent multiple sounds. For example, using the International Phonetic Alphabet, the word mop is spelled “mɒp,” while the word through is spelled “θru,” and cash would be spelled “kæ∫.” It takes a while to get comfortable using the International Phonetic Alphabet. You might well wonder why we do not switch to the International Phonetic Alphabet, as each word’s spelling uniquely describes how that word is pronounced, and this might make spelling easier. This is true, except that it would mean that all homophones (e.g., bear and bare) would be spelled the same, and the meaning would depend on context. As it is, our writing system is made more difficult by the complex correspondence between symbol and sound, but also less difficult by this disambigua- tion of words that have multiple meanings. Using the International Phonetic Alphabet would also mean that regional differences in pronunciation would require different spellings. For example, the word schedule would need to be spelled differently in the United States and the United Kingdom, as Americans and Britons use different pronunciations for this word. It would also require that spellings vary as a function of region within any particular English-speaking country. As Harvard linguist Steven Pinker (1994) noted, for native Bostonians, the words orphan and often would be spelled identically using the International Phonetic Alphabet. Thus, although it may be appealing to come up with a completely phonetic spelling system, it is also not without its problems. TEST YOUR KNOWLEDGE 1. Describe the pathway of air as it moves from the lungs and out of the mouth in making speech sounds. 2. Compare and contrast phonemes and the role they play in speech with letters and the role they play in writing. VARIABILITY IN THE ACOUSTICS OF PHONEMES 12.2 Sketch the mechanisms our speech perception system uses to extract coherent speech from the speech signal. When we hear a speaker utter the expression “The boy was resist- ing arrest,” we must engage in some complex auditory processing to go from the sound signal to perceiving phonemes to under- standing the sentence. For example, could this sentence really be “The boy was resisting a rest,” with far less serious implications? In this case, listeners must use top-down processing using con- textual information to parse the words in a sentence, that is, to break the ongoing stream of language into separate words. If the sentence is said at a day care center for 5-year-olds, we might gravitate toward the latter interpretation. If it is said by a police officer during a trial, we would opt for the first interpretation. Note that for this sentence, the manner in which most speakers of English would say it aloud gives us no clues. In this case, only the context can disambiguate the speech (Figure 12.6). We will see in © iStockphoto.com /boggy22 FIGURE 12.6 Parsing Word Boundaries The difference between “resisting arrest” and “resisting a rest.” © iStockphoto.com /Cylonphoto 348 Sensation and Perception this section that speech perception requires a number of “tricks” that help the auditory system make fast and accurate interpretations of the speech signal. Coarticulation Speech perception must take place at very fast rates, as the average speaker can talk at the rate of four words per second. Or, to phrase it another way, the average speaker can produce about 15 phonemes per second. To produce this rapid speech, our articulators (e.g., the parts of our mouth, vocal tract) must do things simultaneously. Indeed, when we speak, we also anticipate what sounds are coming next. In this way, we will say the phoneme “b” in the word bat in a slightly different way than we say the phoneme “b” in bet, because a different vowel sound follows the “b.” Think about the position of your mouth when you say bat—it is more rounded right at the start than when you say bet. Similarly, the “a” sound in bat is said slightly differently than the “a” sound in back. This phenomenon is referred to as coarticulation. We can define coarticulation as the phenomenon in which one phoneme affects the acoustic properties of subsequent phonemes (ISLE 12.1). What this means is that the consonant and vowel sounds in “ba” overlap when they are spoken. It also means that the “a” sound is influencing how the “b” is being said. This is a reciprocal process, with the vowel influencing the consonant and the conso- nant influencing the vowel. We can see this pattern of coarticulation in sound spec- trograms of people speaking different consonant–vowel pairs. A sound spectrogram is a way of plotting sounds to help understand complex acoustic stimuli. Time is on the x-axis and the different frequencies in the sound are plotted on the y-axis. Finally, the darker the portion of the graph is, the more intense the frequency is. Thus, you can see how the frequencies change in intensity and which frequencies are present over time. The physical acoustics of “ba” are different from those of “be.” This is illustrated graph- ically in Figure 12.7. In Figure 12.7, you can see that the spectrogram looks different right at the beginning for the two different sounds, which both start with the letter b. Coarticulation is an acoustic reality. The acoustic stimuli of “bah,” “bee,” and “boo” are different, as you can see in Figure 12.7. However, as perceivers of speech, we do not hear the differences in the “b” sounds among “bah,” “bee,” and “boo.” All we hear is the phoneme “b.” Thus, even though the signal varies, what we hear is a constant. We encountered the concept of perceptual constancy in our discussion of visual perception. We discussed how we see the same color regardless of the wavelength of ISLE 12.1 Coarticulation 6000 0 0 Time (s) F re q u e n cy ( H z) 0.4944 6000 0 0 Time (s) F re q u e n cy ( H z) 0.542 6000 0 0 Time (s) F re q u e n cy ( H z) 0.6443 FIGURE 12.7 Coarticulation in the Pronunciation of Consonants These spectrograms show the formants for the sounds “bah,” “bee,” and “boo.” You can see that the highest frequency formant is highest for “boo” relative to “bah” and “bee.” This is a consequence of coarticulation. Coarticulation: the phenomenon in which one phoneme affects the acoustic properties of subsequent phonemes 349 Chapter 12: Speech Perception illumination in color constancy, and that objects appeared the same size regardless of retinal size. Our perception of coarticulation is an example of an auditory con- stancy, similar to the concepts of constancy with respect to vision. Because the various “b” sounds refer to the same phoneme, we perceive them as being identical. To reiterate, perception of coarticulation is that two or more different physical sounds are heard as identical phonemes. What this means for our speech perception systems is that several different sounds must be grouped together to represent a particular phoneme, depending on context. This happens early in auditory processing, such that we do not perceive a difference between the “b” in bet and the “b” in bat. This can sometimes be a problem when learn- ing a new language. For example, Japanese speakers group together what we English speakers hear as “l” and “r” into a single phoneme. Thus, when learning English, words such as rail and lair are difficult to distinguish because the two words sound identical to a native Japanese speaker. It is not simply a matter of not being able to pronounce them. Essentially, native Japanese speakers must redefine their phoneme groupings in order to perceive ours. Similarly, there are sounds in other languages that English perceivers do not hear. Nonetheless, our speech perception system groups sounds together, such as the “b” in bet and bat, and the difference between these sounds implicitly allows us to anticipate the sounds that follow each use of the phoneme. Categorical Perception As we have seen, being able to distinguish phonemes is critical to speech perception. Thus, our auditory system uses a number of processes to help us hear the right phoneme. Of course, phonemes represent categories but are made from a range of different physi- cal sounds. To help us with phoneme perception, perceptual variation is clumped using categorical perception. Categorical perception refers to our perception of different acoustic stimuli as being identical phonemes up to a point at which our perception suddenly shifts to perceive another phoneme. That is, we do not hear the variation in the signal—we hear only the phoneme. And then, at a certain point, the variation crosses a boundary, and we hear a different phoneme altogether. Categorical perception exists in other sensory modalities as well. We see blue across a wide range of the visual spectrum, but then blue abruptly becomes green at about 480 nm. Recall the Bornstein et al. (1976) study examining color perception in infants in Chapter 6. It was taking advantage of categorical perception. In speech, we may hear the “r” sound across a wide range of acoustics, but it then abruptly becomes “l” at a certain point. In the case of “l” and “r,” it is the tongue’s position that determines which sound we are making. Sounds with intermediate tongue positions will be heard as either “l” or “r,” and we seldom find these sounds strange. With respect to phonemes, we hear a particular sound, such as “t,” across a range of acoustic stimuli. Then, at a particular point along an acoustic dimension, our perception shifts, and we hear the sound as a different phoneme (e.g., “d”). This is illustrated nicely with a phenomenon called voicing-onset time. Voicing-onset time refers to the time difference between the first sound of the phoneme and the movement of the vocal cords. Variation in voicing-onset time allows us to tell the difference between voiced and unvoiced consonants. The movement of the vocal cords is called voicing. Consider the difference between the pho- nemes “s” and “z.” The phonemes “s” and “z” are similar with respect to the position of the mouth, but “s” is unvoiced, and “z” is voiced. Try saying each consonant while placing your hand on your Adam’s apple. You can feel the vibration for the “z” but not for the “s.” That is, the vocal cords do not vibrate when we say “s,” but they do when we say “z.” The conso- nants “t” and “d” are also a pair of consonants formed the same way but the “t” is unvoiced and the “d” is voiced. In the sound “ta,” our vocal cords vibrate 74 ms after the burst of Categorical perception: the perception of different acoustic stimuli as being identical phonemes up to a point at which perception flips to perceive another phoneme Voicing-onset time: the production of certain consonants (called stop consonants) in which there is a difference between the first sound of the phoneme and the movement of the vocal cords (called voicing) 350 Sensation and Perception sound for the “t,” corresponding to the “a” part of the sound. However, when we say the voiced “da,” there is vocal cord vibration within 5 ms of the start of sound. This is illus- trated in Figure 12.8. You can see an example on ISLE 12.2. In these sound spectrograms, you can see voicing by vertical striations (light and dark areas) going horizontally through the figure. In normal speech, in words such as “ta-ta” or “da-da,” for example, the voicing will be different, with “ta” not showing voicing until much later than “da.” But what happens when we change the natural voicing? This can be done easily in the lab with sound-processing software. In essence, in the lab, we can create a hybrid sound—something intermediate between the “t” and the “d.” Thus, the question becomes, what do we hear when we hear a “t/d”-like sound with voicing half- way in between “ta” and “da”? This is exactly what Eimas and Corbit (1973) did in a now classic experiment on categorical perception. They presented listeners with sounds in which the voicing varied 6000 Beginning of sound (a) Beginning of voicing 0 0 Time (s) F re q u e n cy ( H z) 0.6443 6000 Beginning of sound (b) Beginning of voicing 0 0 Time (s) F re q u e n cy ( H z) 0.5584 FIGURE 12.8 Voicing-Onset Times in Voiceless and Voiced Consonants (a) The sound spectrogram for “ta.” (b) The sound spectrogram for “da.” ISLE 12.2 Voicing-Onset Time 100 (a) “Same” “Different” 80 /da/ /ta/ 60 40 P e rc e n ta g e / d a / re sp o n se s 20 0 0 20 40 Voicing-onset time (msec) 60 80 0 5 10 15 20 25 30 Voicing-onset time (msec) /p a / re sp o n se s (% ) /b a / re sp o n se s (% ) 35 40 45 50 60 0 10 20 30 40 50 60 70 80 90 1000 90 80 70 60 50 40 30 20 10 100 Phonemic boundary 55 (b) FIGURE 12.9 Categorical Perception In an experiment, Eimas and Corbit (1973) asked participants to classify sounds as either “ta” or “da.” They then presented participants with sounds with varying onsets of voicing. They found (a) that almost all sounds with voicing onset lower than 35 ms were classified as “da,” but almost all sounds with voicing onset over 35 ms were classified as “ta.” (b) The 30-ms phonemic boundary for “p” and “b.” 351 Chapter 12: Speech Perception from 0 ms up to 80 ms for “ta” and “da.” At 0 ms, listeners heard the sound as “da,” and at 80 ms, listeners heard the sound as “ta.” Eimas and Corbit were interested in what listeners would hear when intermediate voicings were presented (e.g., 20-, 40-, and 60-ms voicing-onset times). What they found was categorical perception. People heard seemingly normal “ta” and “da” sounds, which abruptly switched from one to the other at a voicing-onset time in between the two. That is, we hear the syllable “da” up to about 35 ms, and then our perception abruptly changes, so that at 40 ms, we hear “ta.” In only a small zone is the stimulus ambiguous, and once the change of perception is made, we do not hear anything odd about these nonnatural sounds. This transition at 35 ms is the phonemic boundary between “t” and “d.” A similar phonemic boundary exists for other voiced–unvoiced pairs, such as “p” and “b,” which transition when the voicing occurs at around 30 ms. Figure 12.9 illustrates these results graphically. Functionally, categorical perception simplifies the determination of what phoneme is being produced and allows us to extract relevant linguistic information from individual and contextual differences in speech. Thus, for example, we could imagine a speaker whose native language is not English, who may voice the vowel after “p” slightly earlier than we do in English. When this speaker says a word such as principal or poet, we hear the “p” sound as a “p” because of categorical perception, which translates a wide range of voicing into one phoneme. The Effect of Vision on Speech Perception and the McGurk Effect Until the advent of telephones in 1876, almost all speech was done within the range of one speaker seeing the other speaker. Yes, people shouted at a distance back then, but speech without vision was rare, to say the least. Even today, outside of telephone conversations, in most cases, we see the people we are speaking to. And with the increased usage of video calling in Skype and FaceTime, we are returning to the time when almost all of our con- versations occur when we can see each other, near and far (see Figure 12.10). We watch as our friends talk to us, we look up toward our teachers when they are address- ing the class (and if we are not looking at them, chances are we are not attending to them), and we watch actors move and talk in movies and television shows. Some of us may spend a great bit of time talking to friends and family on the telephone, particularly if we live some dis- tance from those friends and family. But here is a phe- nomenon you may notice when speaking to a person you do not know via telephone. It is often difficult to distin- guish unfamiliar words over the telephone. A customer representative tells you his name, but you cannot tell if it is Denny or Benny. This is less likely to happen in person because of the presence of visual cues we lack when convers- ing on the phone. In some instances, visual cues can greatly influence our perception of speech. For example, “dubbing” movies requires great skill. If the movements of the mouth do not match up with the sound signal, it can greatly decrease the realism of the movie. Nonetheless, as experimental psychologists, we can ask the question, How much does vision influence speech perception? A very compelling demonstration of the influence of vision on speech perception was developed by McGurk and MacDonald (1976) and has come to be known as the McGurk effect. To demonstrate the McGurk effect, participants are shown a video of a person’s mouth saying monosyllabic sounds, such as “ba,” “da,” and “tha.” However, the audio © iStockphoto.com /Raw pixel FIGURE 12.10 Video Call Chatting Communication McGurk effect: a phenomenon in which vision influences the sounds people report hearing 352 Sensation and Perception component does not always match what the speaker’s mouth was saying when the syl- lables were recorded. That is, the video has been dubbed, and the sound track is simply “ba” over and over again. To be clear, the sound is “ba,” but the movement of the mouth may signal “ga.” The question McGurk and MacDonald posed was what would the par- ticipants hear? Would they hear what they heard or what they saw? Thus, the question McGurk and MacDonald were interested in was whether people would hear the sound track or what the mouth was doing. Strangely, but compellingly, when participants watch the video, they perceive the sounds as being differ- ent both from what they actually hear and what phoneme the mouth makes that they are watching. In Figure 12.11, the observer is watching a mouth say “ga,” but the sound track is “ba.” However, the observer’s perceptual experience is that of hearing “da,” a perception different from the actual sound. To really understand the McGurk effect, though, you have to hear it and see it for yourself. You can see an example on ISLE 12.3. Watch the video once and think about what you hear. Then play the video again and keep your eyes closed. You will be sur- prised to find out that initially, you heard what you saw. The McGurk effect is a very compelling illusion that affects virtually all people. The McGurk effect is surprisingly robust, given that it is fundamentally an error in perception. It works when participants can see only the mouth and when they can see the entire face saying the syllables. It works with people of all ages and lan- guage backgrounds, and it works on faces so degraded that they are scarcely recognizable as faces. Indeed, one study found that there was a McGurk effect for touch (Fowler & Dekle, 1991). In this study, partic- ipants felt a mouth saying one set of syllables, while a sound track played another set. Fowler and Dekle found that the participants heard what they felt. Szycik, Stadler, Tempelmann, and Münte (2012) examined the McGurk effect while participants were undergoing functional mag- netic resonance imaging (fMRI). They found an area of the brain that was active during McGurk effect illusions, that is, when there was a mismatch between audition and vision. This area was not active when participants only heard the sounds or when the sounds and visuals were compatible. This integrative area that drives the McGurk effect seems to rest along the superior temporal sulcus at the top of the temporal lobe. Top-Down Processing and Speech Perception Think about listening to someone speaking in English (presumably a language you are fluent in, if you are reading this text) and about listening to someone speak in a language with which you are not familiar (e.g., Telugu, Yalunka, or Sorani). When listening to English, we hear distinct words, pauses in between words, and then more words again (ISLE 12.4). When listening to an unfamiliar language, it often sounds like a continuous cascade of sound. In a language you are learning, there may be a flip between hearing a continuous stream of speech and hearing breaks between words as your experience increases. In some cases, you may be able to parse the speech correctly into words, but in other cases, you may not. Why does this perception occur? It turns out that speech ISLE 12.3 McGurk Effect Sound from monitor: “ba-ba” Speaker’s lips: “ga-ga” Observer hears: “da-da” FIGURE 12.11 The McGurk Effect An observer is watching a mouth. The mouth is saying “ga,” but the sound track plays the syllable “ba.” The observer does not hear what the actual sound is. Rather, the observer hears a “da” sound. 353 Chapter 12: Speech Perception perception is dependent on a number of top-down processes. That is, knowledge about language influences how we perceive speech. For example, because we know where the boundaries between English words are, we are likely to hear pauses between individual English words. Earlier we pointed out the difference between “The boy was resisting arrest,” and “The boy was resisting a rest.” To hear the difference between these two sentences, we must use top-down processes to determine where the word boundary goes between a and rest. Because we may not be familiar with Telugu (a language spoken in India), we do not hear these word boundaries when we are listening to a person speaking Telugu. To be more specific, evidence for top-down processing in speech perception comes from knowledge of specific combinations of pho- nemes within a language and knowledge of the context of speech. For example, we know in English that “s” sounds may be followed by “p” sounds, but not by “f” sounds. Thus, if we hear someone say “sfecial,” we may assume that what he meant to say was “special.” Or, if the context is different and we know that the person is talking informally about a man or “fellow,” the person may have said “this fella,” but we heard the “s” in “this” run into the “f” in “fella” (see Figure 12.12). Word segmentation is the ability of speakers of a lan- guage to correctly perceive boundaries between words. This means that we use knowledge of our language to draw boundaries as to where one word ends and the next word begins. Consider the following sentences: How to wreck a nice peach. How to recognize speech. In these sentences, we must determine, depending on context, whether the speaker is intending to say three words, wreck a nice, or simply one, recognize. Examining a sound spectrogram will reveal no differences in the pauses between the syllables of the three words and the one word. As a fluent speaker of English, one immediately uses context to determine which of these two utterances is intended. If you are in sensation and perception class at 8 a.m., you are likely to opt for the second sentence, but if you are at a juice bar, and the person making your smoothie accidentally drops a piece of fruit on the floor, you might think the word boundaries correspond to the first sentence. If you are not a fluent speaker of English, you will probably not be able to distinguish these utterances, and therefore not even know how many words the speaker intended to say. Context, which is provided by existing knowl- edge, is therefore an important aspect of speech perception. The Phonemic Restoration Effect We have seen with the McGurk effect that it is possible to induce an illusion of one speech sound even though the actual sound in the environment is another speech sound. In this case, processing of the visual information overrides the input from the cochlea. In the phonemic restoration effect, top-down processing of what one expects to hear overrides input from the cochlea. Like the McGurk effect, phonemic restoration is a very strong subjective effect, and we encourage all readers to go to ISLE 12.5 and listen to it. The phonemic restoration effect refers to an illusion that illustrates the importance of top-down processing for speech perception. ISLE 12.4 Familiar vs. Unfamiliar Languages Wherearethe s i l e n c e s be t w een wo rd s FIGURE 12.12 Parsing Word Boundaries The physical boundaries in sound among syllables do not always correspond to the boundaries between one word and the next. Word segmentation: the ability of speakers of a language to correctly perceive boundaries between words Phonemic restoration effect: an illusion in which participants hear sounds that are masked by white noise, but context makes the missing sound apparent 354 Sensation and Perception In the phonemic restoration effect, an experimenter uses a computer to delete or mask a particular sound in a sentence in which the context clearly indicates what the missing sound should be (Warren, 1970). The experimenter then asks the participant what she just heard. For example, the sentence might be British viewers flocked to the opening **ight performance of Doctor Who. In this sentence, the “n” sound in night has been replaced by white noise, represented here by the double asterisk. Thus, the “n” sound is not actually present, but listeners are asked what they heard. Indeed, listeners report hearing the word night, complete with the “n” sound that is not physically present. It is not merely that they infer that the word must be night; partici- pants actually hear the missing sound. To sample the phonemic restoration effect for yourself, go to ISLE 12.5. What is particularly striking about the phonemic restoration effect is that it works even when the context of the missing sound occurs after the missing sound. Consider the following sentence: It was found that the **eel was on the axle. In this sentence, the context clearly suggests that the missing sound is “wh,” as in wheel. However, at the time the participant hears “**eel,” he has not yet heard the context. We do not learn about the axle until after the missing sound. Nonetheless, when participants heard this sentence (actually used in Warren’s research), they reported hearing the missing “wh” sound. This was not a fluke of this particular sentence. Warren found the same thing when partici- pants listened to this sentence: It was found that the **eel was on the orange. Note that here, the sentence is identical to the previous sentence with the exception of the last word. As in the “wheel” sentence, the context occurs about a third of a second after the missing sound. Nonetheless, here too, participants reported hearing the correct word, in this case peel. For this reason, the phonemic restoration effect has intrigued philosophers as well as psychologists, because speech perception essentially goes back in time and supplies a perceptual experience after the stimulus has been past for almost half a second. It speaks to the strong power of expectations, or what we call here top-down processing, to influence percep- tual experience, even when that experience occurred a split second earlier. Some recent research suggests strongly that the phonemic restoration effect occurs early in information processing, certainly before conscious control, despite the nature of the top-down processing. Mattys, Barden, and Samuel (2014) showed that the per- ception of the missing sound became more pronounced as a secondary task became more difficult. That is, when participants had to devote more attention to a secondary visual task they were doing in addition to the phonemic listening task, the illusion of hearing the missing sound grew stronger. Thus, the phonemic restoration effect must occur at a preattention stage in speech perception processing. This emphasizes how critical the understanding of context is to understanding fluent speech. We can ask what neural mechanisms allow us to unconsciously infer the missing sound even when it occurs before the context is supplied. Sunami et al. (2013) used magnetoencephalography to examine the neurological correlates of the phonemic res- toration effect. They found that areas within the auditory cortex in the temporal lobe were involved with the phonemic restoration effect, presumably the areas responsible for the “hearing” of the missing sound. But just prior to that activation, they found areas in the left prefrontal lobe that responded more strongly to context for missing ISLE 12.5 Phonemic Restoration Effect 355 Chapter 12: Speech Perception sounds than for sounds that were actually present. It is likely that this area of the brain houses the processes that make the top-down inference about context and then what the sound must be. It then transmits this information to the temporal lobe, which “hears” the missing sound. TEST YOUR KNOWLEDGE 1. How does our speech production system create different consonant sounds? 2. Create a diagram to illustrate how categorical perception works, and explain its role in speech perception. THEORIES OF SPEECH PERCEPTION 12.3 Contrast the different types of theories of speech perception, noting their similarities and differences. What should be clear from the discussion so far is that our sensory systems engage in complex processing to transform a complex speech signal into meaning, and they do so amazingly quickly. We have introduced a number of features the human speech per- ception system uses to allow this rapid processing. We attend to coarticulation, that is, that the sound being heard in the present informs us as to what to expect in the future. Our perceptual systems also benefit from categorical perception, which allows us to quickly distinguish among similar phonemes. Furthermore, we use the visual signal of the movements of the mouth to help us understand speech. And we fill in missing sounds when the context suggests what they should be. But are there overall rules that guide speech perception? That is, are there theories that can explain the overall process of speech perception? Such theories can be divided into two general classes. First, some theorists argue that speech is no different than any other sound and that we use the same mechanism with speech that we do with other sounds. These theories are called general-mechanism theories. By contrast, other theorists argue that because of the importance of language to humans, special mechanisms have evolved that are specific to speech and are not used in the processing of other kinds of sound. Not surprisingly, these theories are called special-mechanism theories (Diehl, Lotto, & Holt, 2004). Special-mechanism theories start with the premise that there is a unique neuro- cognitive system for speech perception that is distinct from other kinds of auditory perception. One influential version of the special mechanism is the motor theory of speech perception (Liberman & Mattingly, 1985). This view contends that we have a special mechanism that allows us to detect speech as unique and then relate the sounds to the presumed speech movements (i.e., talking) of the speaker (Galantucci, Fowler, & Turvey, 2006). The goal of speech perception is to infer the movements of the speaker’s mouth. This may sound crazy, but it was developed with the observation in mind that there is seemingly nothing constant about the auditory signal—remember that “b” is a different sound depending on whether it is followed by an “ah” sound or an “eh” sound. In the motor theory view, what is constant across individual phonemes is the speech articulations that produce that sound. Evidence for this view certainly comes from the McGurk effect, in which the speech sound we perceive is what we are seeing, not what is coming in on the sound track. We will see another way to look at special mechanisms later when we discuss how different areas of brain damage influence dif- ferent aspects of speech processing, including speech perception. General-mechanism theories: theories of speech perception that claim that the mechanisms for speech perception are the same as the mechanisms for auditory perception in general Special-mechanism theories: theories of speech perception that claim that the mechanisms for speech perception are distinct and unique relative to the mechanisms for auditory perception in general Motor theory of speech perception: the theory that the core of speech perception is that the system infers the sound from the movements of the vocal tract 356 Sensation and Perception In contrast, the general-mechanism approach argues that speech perception occurs through the same neuro- cognitive processes as other forms of auditory perception. In this view, what makes speech perception unique is its importance, which is learned rather than based on an innate neural mechanism. In terms of theories, scientists tend to prefer a simpler theory unless there is compelling evidence to support a more complex theory. In this case, the general-mechanism theory is less complex than the special-mechanism theory because the general-mechanism theory requires only one neurocognitive system. Thus, this theory should be considered the default until pro- ponents of the special-mechanism theory can definitively demonstrate that speech perception requires its own unique mechanism. One general-mechanism approach includes the view that speech perception involves top-down mechanisms of categorization. Thus, any sound is analyzed by processes that categorize sound in general. When we hear speech, we immediately classify it as such. If we hear the creaking of a door, we do not invest any meaning into that sound other than that someone or something is pushing on that door. If the sound is something like “meow,” we classify it as a nonspeech sound, although, in this case, one that might contain information (Figure 12.13). Any cat owner knows the difference between her cat’s meow that means “feed me” and the one that means “open the door.” However, when we hear a stimulus that sounds like “Open the door, please” (I’m not a cat, so of course I say “please”), we immediately classify many features of the utterance, such as the speaker’s gender, regional dialect, age, emotional state, and identity. It is this act of classification that represents what makes speech perception unique (Holt & Lotto, 2010). Holt and Lotto (2010) pointed out that in many cases, nonspeech sounds can influence our perception of speech sounds, consistent with the gener- al-mechanism approach. For example, in one study (Holt, 2005), a speech sound such as “ga” or “da” was preceded by a series of 21 nonspeech sounds. Each series created a different history of sound perception intended to influence the perception of the speech sounds presented at the end. If speech perception occurs by a unique and special mechanism, the pattern of preceding nonspeech sounds should have no effect on the perception of speech. However, if the general-mechanism approach is correct, then there is the possibility that nonspeech sounds influence speech perception. And this is what Holt found. The pattern of the 21 nonspeech sounds influenced the likelihood of hearing the speech sound as “ga” or “da,” consistent with the general-mechanism view (Figure 12.14). Because of studies such as Holt’s (2005) and others, most researchers today think of speech perception as a form of auditory perception in general. The tricky part, however, is still how we extract speech signals, such as phonemes, from complex and variable stimuli. Thus, many speech perception researchers have moved away from the special-mechanism model but still think the goal of speech perception is to connect sound to the source of that sound, namely, the movements of the vocal tract (Fowler & Thompson, 2010). Fowler’s view is a direct perception view (e.g., Gibson, 1979). She argues that the goal of percep- tion is to guide us through the environment. In the case of speech perception, Fowler argues that it is the movement of the vocal tract that is the invariant © iS to ck ph ot o. co m /N ev en a1 98 7 FIGURE 12.13 Do We Know What Cats Are Saying? Cat owners know the messages their cats are sending when they listen to their meows. Cats use a different meow to indicate “feed me” than they do for “scratch my ears.” 100 80 60 40 20 0 1 2 3 4 5 Consonant–vowel stimulus P e rc e n ta g e o f “g a ” re sp o n se s 6 7 8 9 High Mean Mid Mean/ Low Variance Low Mean Mid Mean/ High Variance FIGURE 12.14 Perception of Speech Is Influenced by Nonspeech Sounds (Holt, 2005) The more repetitions of the nonword stimulus, the less likely the test stimulus was heard as a speech sound. 357 Chapter 12: Speech Perception in the word, and that the perceptual apparatus is tuned to that invariant. In other words, our speech perception is really about figuring out what the vocal tracts of others are doing. Thus, it is no wonder that our speech perception is sometimes fooled by sounds, if our goal is to infer how somebody’s vocal tract works rather than to perceive sounds per se. The Development of Phoneme Perception Infants are born into the world with a remarkable aptitude for acquiring language but no genetic penchant for learning one language or another. The language learned is, of course, determined by the linguistic environment the infant finds himself in. An infant raised by an English-speaking mother will learn English, and an infant raised by a Sorani-speaking mother will learn Sorani (Kurdish). However, languages vary in the phonemes they use, and even when they use the same phonemes, there may be subtle differences among them. For example, the English “p” sound is more explosive than the Spanish “p” sound. During the first year of life, infants are engaged in an intensive learning of the rules of speech perception in the language or languages that surround them. In the United States, monolingualism is the norm, but in many other countries, children are bilingual or multilingual from a very early age. During the first 6 months of life, an infant can distinguish among all of the pho- nemes used in a language and can discern subtle differences in sound among equivalent phonemes across languages (Harley, 2014). We may think of infants at this age as not being linguistic in nature, as they cannot speak yet and do not understand meaning, but these infants are paying very close attention to sound. Indeed, they are paying closer attention to sound than are the adults around them. Consider the difference between the English “p” and the Spanish “p.” When we hear a person who is a native speaker of Spanish speak in English, we may note that she has an accent, but when we listen to her speech, we are attending to meaning, not subtle differences in sound. The difference between saying “Harry Potter” with an explosive English “p” or “Harry Potter” with a soft Spanish “p” is irrelevant. Either way, we are talking about a fictional wizard. However, young infants are attending to just these differences (Harley, 2014). And, just in case you do not know, Harry Potter should not go back to Hogwarts. Consider a study done by Janet Werker and her colleagues (e.g., Maurer & Werker, 2014). They compared infants and adults in their perceptions of the two ways of saying the “t” sound in Hindi, a language widely spoken in India and throughout the Indian diaspora. In Hindi, there is a “t” sound spoken with the tongue at the back of the mouth, and a “t” spoken with the tongue at the front of the mouth, as it is in English. Because the sounds may be associated with different meanings in Hindi, Hindi adults recognize the difference between these two sounds. To English-speaking adults, how- ever, they both sound like “t,” and we do not hear the difference. In one of her seminal studies, Werker looked at whether infants being raised in English-speaking homes would hear the sound difference between the two Hindi “t” sounds (see Werker, 2012). Werker found that at 6 months of age, these infants were able to distinguish the two sounds with nearly 100% accuracy. However, by the time these infants were 10 months old, their accuracy had dropped to about 20%. In con- trast, Indian babies (being raised in a Hindi-speaking environment) maintained this 100% accuracy at 10 months and beyond. Thus, because the sound difference is rel- evant in Hindi, Hindi babies maintain the distinction, but because it is irrelevant in English, English babies no longer perceive the difference. These results illustrate something that Werker calls perceptual narrowing (Werker & Gervain, 2013). Perceptual narrowing refers to the developmental process whereby regularly experienced phonemes are homed in on, as well as the simultaneous diminish- ing of the ability to discriminate unfamiliar phonemes. That is, as infants get older, they Perceptual narrowing: the developmental process whereby regularly experienced phonemes are homed in on, with simultaneous diminishing of the ability to discriminate unfamiliar phonemes 358 Sensation and Perception focus their attention on stimuli that are relevant to them, rather than attending to all stimuli out there. This is advantageous in learning one’s first language, though it makes it more difficult to acquire a native-sounding accent in a second language later on. We have focused here on the experiments comparing Hindi-language infants and English- language infants, but one can see these perceptual narrowing differences across many language pairs (Harley, 2014). Because speech signals are so complex, it benefits young children (and, later, adults) to ignore irrelevant differences in sound and focus only on those sound differences that are meaningful in the language that they speak. Normal speech development, as mentioned earlier, takes place in an environment rich in language. But what happens if the child grows up without that stimulation? Tragically, there have been a few cases of children that were raised in extreme social isolation that removed the child from this essential linguistic environment. One famous case is that of a child who is called Genie (Curtiss, 2014). After having suffered tre- mendous abuse under horrific circumstances, she was discovered at the age of 13 and was taken from her home. She had grown up socially isolated and had minimal con- tact with her mother and none with her father or other family members. The degree and duration of isolation were extreme even relative to most other cases (Fromkin, Krashen, Curtiss, Rigler, & Rigler, 1974). Without going into details, the circumstances were unspeakable. However, she survived. She was unable to speak when discovered, though there is some evidence that she might have begun normal language develop- ment within her first 20 months. Perhaps some of the perceptual narrowing mentioned by Werker above might have taken place. Still, she could not speak or understand language at first. Her communications at first can be described as whispered grunts. Still, over time, Genie has shown a remarkable ability to learn to speak and, particu- larly, to understand speech. She learned to understand the difference between singular and plural nouns, negatives, possessives, and a number of prepositions, particularly in regard to perception and comprehension (Fromkin et al., 1974). Although she never likely developed adult language, her development is remarkable, nonetheless (Curtiss, 2014). Early development is extremely important for language, but it appears that some speech learning abilities can persist into later life. You can see some video of Genie in ISLE 12.6. TEST YOUR KNOWLEDGE 1. What are the general-mechanism and the special-mechanism theories of speech perception? What differences do they anticipate? 2. What is perceptual narrowing? What does it do for a developing child? Can you think of a phenomenon later in life in which perceptual narrowing leads to difficulties? SPEECH PERCEPTION AND THE BRAIN 12.4 Diagram the areas of the brain involved in speech perception and what happens when they are damaged. Understanding language is an extremely important aspect of being a human being. Therefore, we should expect to see complex neural systems involved in the processing of the human speech signal. And this is exactly what we find in the human brain. Although we focus on perceptual aspects of language, equally important to language is producing ISLE 12.6 Genie Broca’s area: an important area in the production of speech, located in the left frontal lobe Wernicke’s area: an important area in speech comprehension, located in the left temporal lobe Aphasia: an impairment in language production or comprehension brought about by neurological damage Broca’s aphasia: a form of aphasia resulting from damage to Broca’s area, causing a deficit in language production 359 Chapter 12: Speech Perception speech. Indeed, there is close integration of the perception and comprehension networks for lan- guage and the brain regions responsible for pro- ducing speech. Two 19th-century neurologists, Pierre Broca and Carl Wernicke, were instrumental in begin- ning our understanding of the neuroanatomy of speech and language (Figure 12.15). For their work, each has an area in the brain named after him: Broca’s area and Wernicke’s area. Interestingly, the patient Broca studied had dam- age in an area different from the one now named after him, but Broca is given credit for being one of the first to associate specific areas of the brain with unique language functions. Wernicke cor- rectly identified the areas for the production of language and for the comprehension of language. For this reason, some anatomists call Broca’s area Broca–Wernicke’s area to honor the bet- ter scientist but still distinguish it from Wernicke’s area proper. We use the term Broca’s area, despite Broca’s own mis- identification of the area. It is also impor- tant to note that Broca’s and Wernicke’s areas are not the only areas in the brain associated with language, and there is even some debate as to the role of these two areas in language function. Broca’s area and Wernicke’s area are depicted in Figure 12.16. Broca’s area is in the left frontal lobe directly in front of the primary motor cor- tex and is an important area in the pro- duction of speech. Wernicke’s area is in the left temporal lobe and is critical in under- standing language. In the vast majority of people, Broca’s and Wernicke’s areas are localized in the left hemisphere, although a very small minority, most of them left- handers, may have these areas localized in the right hemisphere. Aphasia is an impairment in language production or comprehension brought about by neurological damage. In Broca’s aphasia, the damage is to Broca’s area of the brain. Broca’s aphasia is characterized by nonfluent speech. However, by and large, speech perception is not affected, and language comprehension is normal. Broca’s aphasics have a halted speech pattern and have difficulty speaking sentences. Their frustration at their inability to speak can sometimes be palpable, as they have an understand- ing of meaning, but cannot translate it into speech. There is also some evidence that Broca’s aphasics have deficits in understanding complex grammar relative to controls, even though their word comprehension shows no such deficit. A video of patients with Broca’s aphasia is available in ISLE 12.7. ISLE 12.7 Broca’s and Wernicke’s Aphasia Paul Fearn/A lam y Stock Photo Sc ie nc e & S oc ie ty P ic tu re L ib ra ry /S SP L/ G et ty Im ag es FIGURE 12.15 Two Early Scientists on Language and the Brain: Pierre Broca (left) and Carl Wernicke (right) FIGURE 12.16 Broca’s and Wernicke’s Areas Broca’s area is in the frontal lobe, and Wernicke’s area is in the temporal lobe. Broca’s area Motor cortex Angular gyrus Primary visual cortex Wernicke’s area Lateral �ssure 360 Sensation and Perception Wernicke’s area is located in areas of the brain associated with the auditory system in the tempo- ral lobe. Anatomically, it is located in the posterior section of the supe- rior temporal gyrus. This area is also known as Brodmann area 22. Damage to Wernicke’s area results in deficits in the comprehension of lan- guage, a condition called Wernicke’s aphasia. Severe Wern icke’s aphasia may result in a complete absence of understanding language. Speech is, by and large, fluent, but it may appear to not make sense to listen- ers, as the patients themselves cannot understand what they are saying. This meaningless speech is some- times called jargon aphasia. Unlike a person with Broca’s aphasia, individ- uals with Wernicke’s aphasia often show a blithe indifference to their disorder and seem unaware of their problems. Videos of patients with Wernicke’s aphasia can be found in ISLE 12.7. Interestingly, although Wernicke’s area clearly emerges from auditory processing parts of the brain, its role in language is comprehension. There are cases in which sign lan- guage speakers have developed brain damage in Wernicke’s area. These sign language speakers have a deficit in understanding sign language equivalent to the deficit in comprehension seen in Wernicke’s aphasia for spoken language (Bellugi, Klima, & Hickok, 2010). Thus, in people whose language is visual–manual instead of auditory, it is areas of the brain associated with the auditory cortex that are involved in inter- preting language. As with any neurological damage, the extent of the brain damage may vary from patient to patient. Thus, a patient with a relatively small lesion in Wernicke’s area may be able to produce sensible speech and understand some elements of others’ speech. However, a patient with a large lesion may not understand speech at all. As you have probably guessed by now, nothing having to do with perception is simple. This is true for the neurological basis for speech perception as well. Although Wernicke’s area is critical for speech perception, there are many other areas of the cortex also involved in this process. This includes auditory perception areas in the temporal lobes as well as monitoring areas in the left prefrontal cortex. Studies with neuroimaging techniques have delineated some of the connections and links between these areas. In particular, many neuroscientists now think that the distinction between “what” and “where” is applicable to speech perception. That is, our auditory cortex has pathways for identifying the “what” of the signal, namely, recognizing meaning Via higher order frontal networks Articulatory network pIFG, PM, anterior insula (left dominant) Spectrotemporal analysis Dorsal STG (bilateral) Phonological network Mid-post STS (bilateral) Conceptual network Widely distributed Combinatorial network aMTG, aITS (left dominant?) Lexical interface pMTG, pITS (weak left-hemisphere bias) Ventral stream Dorsal stream Input from other sensory modalities Sensorimotor interface Parietal–temporal Spt (left dominant) (b) (a) FIGURE 12.17 The Dual-Stream Model of Speech Perception (Hickok & Poeppel, 2007) The dual-stream model of speech perception involves a ventral pathway that recognizes speech as speech and a dorsal stream that links that speech to the movements of individual speakers. Abbreviations refer to regions of the brain. The regions are color coded so that the color of the box matches the brain region. For the intensely curious: aITS, anterior inferior temporal sulcus; aMTG, anterior middle temporal gyrus; pIFG, posterior inferior frontal gyrus; PM, premotor cortex; pMTG, posterior middle temporal gyrus; Spt, parietal lobe temporal lobe boundary; STG, superior temporal gyrus; STS, superior temporal sulcus. Wernicke’s aphasia: a form of aphasia resulting from damage to Wernicke’s area, causing a deficit in language comprehension 361 Chapter 12: Speech Perception in speech, and a “where” system, which processes where the speech signal is com- ing from (Hickok & Poeppel, 2007). Figure 12.17 shows these areas associ- ated with the two streams. Some other neuroscience evidence is relevant to the topic of whether speech perception constitutes a special mecha- nism or is best thought of as a function of a general auditory perception mechanism. This research concerns the discovery of a voice area in the superior temporal sulcus. Research shows that this area becomes more active in response to human voices than it does to nonspeech sounds (Belin, Bestelmeyer, Latinus, & Watson, 2011). Indeed, fMRI studies show that an area Belin et al. (2011) called the temporal voice area, located in the superior temporal sulcus, shows greater responses to voices than to natural nonvoice sounds or to unidentifiable noise (Figure 12.18). One of the questions that has always fascinated neuroscientists and laypersons alike is whether we can tell what a person is thinking on the basis of the observable activity of his brain. Many science fiction movies and novels involve fancy electronic equipment that can determine exactly what someone is thinking. With the advent of neuroimaging, perhaps we can ask whether a neuroscientist watching the activity of your brain in real time with fMRI might be able to know that you are thinking about the time you were looking over the rim at the Grand Canyon, or about how much Descartes’s religious beliefs informed his philosophy. Or perhaps you were thinking about something you would prefer the neuroscientist not know about and that we will not mention here. This kind of research is still in its infancy, but some fascinating research has been able to infer what words a person is listening to based on the activity in her auditory cortex (Pasley et al., 2012). In this unique and fascinating study, Pasley et al. (2012) recruited patients who were soon to undergo surgery on their temporal lobes as a treatment for epilepsy or brain tumors. All of the participants had normal language abilities. Because of the impending neurosurgery, the patients had been fitted with intracranial electrodes, that is, sensitive electrodes fitted on the temporal lobe underneath the skull, thus allowing fine spatial and temporal resolution of ongoing human brain activity. While these electrodes were actively recording brain activity, the patients heard recordings of words and sentences from a variety of speakers. This is interesting enough as it is, but what Pasley et al. did then was feed the pattern of brain activity into a computer programmed to decode what those words and sentences were. This program did not originally know what the stimuli were, so it allowed Pasley et al. to determine if it was possible to infer what the speech areas of the brain were listening to by observing activity in the auditory cortex alone. In essence, their program serves as a speech decoder, taking the brain’s output to re-create the input. So here is the goal of the algorithm built into the computer program: Based only on the pattern of activity seen in the temporal lobe, can the program extract the orig- inal auditory signal? In this way, an experimenter will know what word was spoken to the patient without actually having heard the word, but only seeing the output FIGURE 12.18 The Temporal Voice Area (Belin et al., 2011) The temporal voice area is located in the superior temporal sulcus, shown here. It responds more to voice than to nonvoice sounds. Voice area: an area located in the superior temporal sulcus that responds to the sound of the human voice, but less so to other stimuli 362 Sensation and Perception of the computer program. Although not perfect—in the best case, it was still less than 60% accurate—the algo- rithm in the computer program was able to re-create the auditory signal on the basis of the pattern of activity in the auditory cortex, and it did so particularly well for words, though less well for sentences. You can see both the process and the results in Figure 12.19. In the words of Pasley et al. (2012), “The results provide insights into higher order neural speech processing and suggest it may be possible to read out intended speech directly from brain activity” (p. 2). We think this is an important and bold step not just in the understanding of speech perception but in the way in which neuroscience is heading. TEST YOUR KNOWLEDGE 1. What is aphasia? Develop a classification system for Broca’s and Wernicke’s aphasia. 2. Describe what we learn about how the brain processes speech from aphasia. Acoustic waveform Time (s) Reconstructed spectrogram Reconstruction model Cortical surface �eld potentials Time e le ctro d e s X tim e la g s Fr e q u e n cy ( kH z) 0 0.2 1 7 2 = x FIGURE 12.19 Reinterpreting Brain Signals (Pasley et al., 2012) This figure shows how the original signal is recorded in the brain and then re-created via the algorithm that interprets the electrical activity in the temporal lobe. EXPLORATION: Hearing Loss and Speech Perception Taylor Monroe is a fictional 45-year-old woman. She has a graduate degree in education and has been teaching high school social studies for 20 years, and she is one of the more popular teachers among the students. Ms. Monroe has been wearing hearing aids to compensate for sensorineural hearing loss since she was in college. Ms. Monroe’s hairstyles tend to cover her ears, so most of her students do not know that she has a hearing impairment. Ms. Monroe uses the latest tech- nology in hearing aids, but she still complains that it is often difficult for her to understand her students’ questions when there is more than one person talking or if there is construc- tion noise outside. Ms. Monroe’s complaint is typical of many hearing-impaired people. Even with amplification, speech comprehension may still be difficult, especially when other sounds are present. Although Ms. Monroe has had her whole life to adjust to her hearing loss, this problem is even more per- vasive in older adults who experience hearing loss late in life. As we discussed in the previous chapter, today’s digital hearing aids are amazing apparatuses that use complex engineering to help people hear (Figure 12.20). As recently as 20 years ago, hearing aids merely amplified sound. But now they have complex programs to help individuals with hearing loss, especially with the problem of speech percep- tion. Nonetheless, the major goal of hearing aids still is the selective amplification of frequencies to compensate for higher thresholds at those frequencies (Kochkin, 2005). FIGURE 12.20 Hearing Aids Hearing aids are small and easily concealable, but they help people with hearing disabilities in tremendous ways. © iS to ck ph ot o. co m /s na pp ho to 363 Chapter 12: Speech Perception Amplification of speech brings sounds for which an indi- vidual is hearing impaired above threshold, so that he can hear them, but the same amplification may also amplify sounds the person can hear normally, resulting in their being too loud. As we know, any voice has multiple har- monics (formants), some of which may not be amplified and some of which may be amplified too much. These frequency components that are too loud introduce a new kind of distortion. Many of the latest tricks in hearing aids are designed to compensate for this aspect of hearing. But because one can never be certain of what a new person’s voice will sound like, in practice, this is very difficult. To summarize the problem, when hearing aids compensate for loss of sensitivity, they may introduce distortion by amplifying irrelevant aspects of the signal. Imagine a speaker saying an unexpected sentence, such as “The Jade Rabbit landed on the moon.” Unless you have been following the Chinese space agency’s moon explo- ration program, this sentence might not make sense to you at first. In the case of our hypothetical teacher, con- text may not lend much of a hand to help Ms. Monroe figure out what her student just said about a jade rab- bit. As we have emphasized in this chapter, speech takes place over time. It takes a small amount of time to say the word rabbit, and coarticulation helps listeners decode sounds. However, if there is noise, either natural or dis- tortion produced by hearing aids, comprehension of the words and sentence may be slowed to a point at which the person cannot make intelligible sense of the words. Amid noise, hearing-impaired individuals have more difficulty following the stream of speech, because they are missing cues that are present when there is less noise or that are always available to those with normal hearing. The noise may interfere with the signal such that syllables within words are less clearly separated in time. Thus, people with hearing loss have difficulty understanding speech in noise (Henry, Turner, & Behrens, 2005). That is, because speech perception is so complex, and the processes must be achieved so quickly, those with hearing loss may not have sufficient time to process speech stimuli, even when they are hearing sounds amplified with their hearing aids. This problem may be even more profound in older adults who develop hearing loss later in life (Anderson, White- Schwoch, Parbery-Clark, & Kraus, 2013). The situation for setting up hearing aids becomes even more complicated when different languages are con- sidered. Different languages have different phoneme structures, leading to the differences in development of phoneme perception that we have already discussed earlier in the chapter. Current methods of programming hearing aids can allow the aids to be adjusted to be sensitive to some of these differences in phonemes. For example, Slavic language speakers need hearing aids programmed to boost frequencies in the 3,000 to 3,500 Hz range when compared with hearing aids for English speakers, whereas Chinese speakers need their hearing aids to boost the amplitude of frequencies more in the range of 2,700 to 3,000 Hz (Chasin, 2011). However, capturing phonemes does not reveal all of the differences between languages that occur at the word and sentence levels. For example, English has a subject-verb- object basic word order for sentences. Other languages, such as French, Spanish, and Tamil, have a subject-object-verb word order. We often reduce the intensity of the speech sounds at the end of our sentences, but if that end word is a verb, and you have hearing loss, then a subject-object- verb word order would cause you to have more trouble understanding what is being said than for languages with word orders like English does. In one study, Chasin (2012) found that for these subject-object-verb word order lan- guages you need to change the programming so that the ends of sentences are selectively amplified relative to lan- guages with other word orders. Hearing aids should also be sensitive to other differences as well, such as “tone,” seen in languages such as Mandarin Chinese, other Chinese languages, and Vietnamese (and also some West African languages). English is a language that forms its phonemes by the structure, not the pitch, of the sound (see Chapter 13 for more on what pitch is). Chinese and Vietnamese use pitch to denote differences in phonemes and therefore differences in meaning. Let us look at vowels. As mentioned earlier, it is the ratio of the formants that plays a key role in determining the phoneme, whether an “a” or an “e,” in English, not the pitch at which the sound is spoken. In Chinese, different pitches can indicate different phonemes. Current hearing aid technology and even cochlear implants do not repro- duce pitch as well as they do consonants. For example, Wong and Wong (2004) found that Cantonese children with cochlear implants had difficulty discriminating the tones they needed to understand their language. Given that hearing aids are of limited help in many lan- guages, there are some general recommendations given to help communicating with hearing-impaired individuals. For example, the speaker can stand close and face-to-face to the hearing-impaired listener. In addition, the speaker should speak slowly and distinctly. It also helps to use gestures and posture to help convey what is being said. Eliminating distracting sounds can also be very important 364 Sensation and Perception to communicating to a hearing-impaired person (Huang & Tang, 2010). To summarize, individuals with hearing impairments choose to wear hearing aids to help them understand speech. However, evidence suggests that some difficul- ties in speech perception are actually introduced by the pattern of amplification and suppression from the hear- ing aids themselves. In essence, the combination of the hearing aid and the impaired cochlea results in a num- ber of ways in which speech perception can be affected adversely. Because the signal going into the auditory nerve is impaired, it requires the auditory cortex to work harder to decode the signal, leading to impaired percep- tion amid noise or if the individual is distracted (Leek & Molis, 2009). The solution is to design hearing aids with faster and more efficient processing with greater sensitivity to dif- ferent languages, words, and sentences. But the exact nature of that processing will require research to deter- mine what features of the human auditory system need tweaking in each individual patient. Although appropriate amplification may address the problem of impaired hear- ing in general, it may not be enough to compensate for the slowed temporal processing and distortion prevalent in people with moderate sensorineural hearing loss (Leek & Molis, 2009). As the science of hearing aids continues to improve, patients can expect to find hearing aids with a greater range of internal programs that adjust the acous- tics of the situation automatically depending on ambient conditions. Thus, in a noisy environment, hearing aids will focus on amplifying sounds in the frequencies commonly heard in speech. In addition, loud sounds will generate a quick attenuation of the amplification of those sounds, such that coughs or squeaking chairs lead to less inter- ference with speech perception. When the person moves to another environment, say a quiet room at home, the hearing aid will adjust to that room and inhibit only some high-frequency sounds, such as the hum of an air con- ditioner or the buzz of a refrigerator, leaving the person maximally able to converse with his or her family. Even with hearing aids, Ms. Monroe may never be able to pick up speech in a noisy classroom as well as her nonimpaired colleagues, but hearing aids are evolving quickly to maxi- mize speech perception. APPLICATION: Hey Siri, or Are You Cortana? Computer Speech Recognition “Hey Cortana, tell me a joke.” “I wondered why the baseball was getting bigger. Then it hit me.” “Alexa, who’s your favorite Beatle?” “John, Paul, George, or Ringo.” Who are Cortana and Alexa? Cortana is the name giv- en to the search function on Windows machines, and Cor- tana has the ability to both respond to verbal commands and “speak” its responses. Alexa is the name used to talk to the Amazon Echo, a home device that can play music, set times, control lights, and perform an increasing number of other home functions, all by speaking to it (Figure 12.21 shows one such device). So the conversations above are possible short conversations one can have with a computer. In recent years, there has been an explosion of devices that allow you to talk to them. In addition to these devices, there is “Siri” for Apple devices, and you can say “OK, Google” to activate speech recognition on Android phones. Originally you had to press a button to start the device listening, but now you can use a wake-up word such as “Hey Siri” to activate the application. In other domains, speech recognition is also being developed. Many cars now have speech recognition systems that allow you to tell your car’s computer where you want to go without taking your eyes off the road. Thus, now, you can talk to your phone, tablet, laptop, or car. Speech recognition can also be found in military applications. We are not far at all from the old Star Trek “Computer, tell me whatever I need to know to further the plot.” Even back then, the creators of Star Trek recognized the need for a wake-up word so that the device knows when you are talking to it, and addressing the computer as “Computer” was their way, though as you see, the actual trend is to give our computers women’s names. As with so much of 365 Chapter 12: Speech Perception perception, the current speech recognition abilities of these devices seem easy, and we will soon take for granted these abilities, but years of research understanding speech per- ception processes lay behind these remarkable new tech- nologies. In fact, research on speech recognition goes back to the 1930s at the remarkable Bell Laboratories (Juang & Rabiner, 2005). Thus, there is a lot of work behind these neat little devices that must solve many of the same issues that our perceptual system does to understand speech. Let us look at a bit of it. As discussed earlier, there is a lot of ambiguity in the speech stimulus. Coarticulation and other features make for a lot of variations. Whatever device is listening to you must be able to decode those ambiguities to cor- rectly interpret what you are saying. The problem is the same for the device as it is for our brain. It is clear that a great deal of progress has been made in understanding our speech, but these devices are certainly not perfect. Words that are not commonly used and those that have multiple pronunciations are often more confusable. This is true for us, but to a much greater degree for speech recognition systems. Try asking your device, “Do I have a dactylion?” You might not know what the word means (it’s an anatomical marker at the tip of the middle finger), but you could repeat the word back. One speech recog- nition system responded with “Do I have a deck tillian?” The last part is not even a word. There are also features of speech that can confuse these systems. For example, if you are speaking along and use some sort of nonspeech sounds, like the ever present “uh,” your speech recognition system might struggle to get what was said correct. Yet, these sorts of nonspeech sounds rarely impede our own understanding; just listen to your professor (Goldwater, Jurafsky, & Manning, 2010). If you watch your voice recognition system as you talk, some of them show their current estimate of what you say as you go along. You can see several guesses for a partic- ular word before it finalizes on what it thinks you said. These systems try to consider linguistic context, just as we do, to disambiguate what is being said by you. One of the ways that these systems work is to use an expectation of word order to figure out your sentence. As was mentioned in the Exploration section of the chapter, English uses a subject-verb-object word order, and speech recognition sys- tems for English expect that word order. So, mixing syntax can confuse these systems much more than us. For example, say to your device, as the Star Wars character Yoda might, “Red the apple is” and see what you get. Cortana recently replied, “Where did the apolis?” Cortana was confused and did not make sense. Our ability to respond to context is much more fluid than these systems that work best with simple, expected grammatical constructions. Another issue for these systems is accents. Although we may struggle with determining what people with second-language accents or regional accents say, mostly we can figure them out, even if we occasionally have to ask for clarification. However, speech recognition systems still struggle with accents despite efforts to improve them (Hansen & Arslen, 1995). So, despite the great advances in trying to figure out the words we are say- ing, there is progress to be made in computer voice recognition when words are pronounced in unusual ways (Abdelaziz & Kolossa, 2016; Schädler, Warzybok, Hochmuth, & Kollmeier, 2015). Again, the issues speech recognition systems face are the same ones we face in understanding language. Think of the phonemic restoration effect. Recall that the phonemic restoration effect occurs when expected words are removed from the speech signal. Our own speech recogni- tion is so efficient that we can remove a significant amount of the speech stimulus and still understand what is being said to us. This ability is vital as we are rarely in a completely quiet environment. Other sounds occur and can cover or mask another person’s speech. Think of a party with loud music and lots of conversations or the situation created in the cocktail FIGURE 12.21 The Amazon Echo Speech Interface Boston G lobe/G etty Im ages 366 Sensation and Perception party effect. We still can follow along with what is being said in our conversation, despite the noise around us. A computer speech recognition system must be able to ultimately perform at the same level in noisy environments, but they are not there yet. Dai and Soon (2013) developed an algorithm that tries to simulate human abilities to extract speech sounds out of other noise. The algorithm tries to examine the temporal and other characteristics of the sound to determine what is speech and what is not speech. However, this aspect of speech recog- nition cannot yet match what any person can do. Speech recognition has advanced greatly in recent years, but there is still progress to be made. Although these systems use computers and complicated mathematics to solve their problems, they are the same problems that our auditory systems have to solve with tissue and nerve fibers—that is, to understand what is being said every day. Knowledge of human speech perception aids these researchers in their work, and it is not too much to expect that the solutions for computer speech recognition will resemble how our brains accomplish the same task. CHAPTER SUMMARY 12.1 Discuss the complex process of speech percep- tion and speech production with regard to the speech stimulus. Speech perception is the most critical aspect of auditory perception for human beings. Perceptual processes begin with a stimulus in the environment. In the case of human speech perception, that stimulus is a voice. The human vocal tract produces both consonant and vowel sounds, which are the bases of the sounds used in human language. Formants are the frequency bands with higher amplitudes among the harmonics of a vowel sound. Each individual vowel sound has a specific pattern of formants. The place of articulation is the point along the vocal tract at which the airflow is constricted. The manner of articulation is how that restriction occurs. Voicing refers to whether the vocal cords are vibrating or not. Phonemes are the basic units of sound in human language. The International Phonetic Alphabet is an alphabetic convention that provides a unique symbol for each and every phoneme in use in human languages. 12.2 Sketch the mechanisms our speech perception system uses to extract coherent speech from the speech signal. A problem for our speech perception systems is to iden- tify what is constant in the speech signal. Our auditory system uses a number of mechanisms to identify pho- nemes and understand speech. Coarticulation is the phenomenon in which one phoneme affects the acoustic properties of subsequent phonemes. Our auditory sys- tems pick up on coarticulation and use it to understand speech. Categorical perception is our perception of dif- ferent acoustic stimuli as being identical phonemes, up to a point at which our perception flips to perceive another phoneme. Voicing-onset time is the production of certain consonants (called stop consonants) in which there is a difference between the first sound of the phoneme and the movement of the vocal cords. The movement of the vocal cords is called voicing. In the McGurk effect, par- ticipants are shown a video of a person’s mouth saying monosyllabic sounds, such as “ba,” “da,” and “tha.” However, the audio component does not always match what the speaker’s mouth was saying when the syllables were recorded. The results show that the visual input affects what people report hearing. Word segmentation is the ability of speakers of a language to correctly perceive boundaries between words. The phonemic restoration effect is an illusion in which participants hear sounds that are masked by white noise, but context makes the missing sounds apparent. 12.3 Contrast the different types of theories of speech perception, noting their similarities and differences. General-mechanism theories of speech perception claim that the mechanisms for speech perception are the same as the mechanisms for auditory perception in general. Special-mechanism theories claim that the mechanisms for speech perception are distinct and unique relative to the mechanisms for auditory perception in general. The motor theory of speech perception argues for a special mechanism that makes attributes about sound from the movements of the vocal tract. 12.4 Diagram the areas of the brain involved in speech perception and what happens when they are damaged. Perceptual narrowing is the developmental process whereby regularly experienced phonemes are homed Chapter 12: Speech Perception 367 in on, with the simultaneous diminishing of the ability to discriminate unfamiliar phonemes. Broca’s area is in the left frontal lobe and is an important area in the produc- tion of speech. Wernicke’s area is located in areas of the brain associated with the auditory system in the tempo- ral lobe. Aphasia is an impairment in language produc- tion or comprehension brought about by neurological damage. Broca’s aphasia is a form of aphasia resulting from damage to Broca’s area. In Broca’s aphasia, the deficit is in language production. Wernicke’s aphasia is a form of aphasia resulting from damage to Wernicke’s area. In Wernicke’s aphasia, the deficit is in language comprehension. The voice area is located in the superior temporal sulcus and responds to the sound of the human voice. In a fascinating study by Pasley et al. (2012), a com- puter program was able to determine what a speaker was saying based only on the activity of the speaker’s auditory cortex. Speech perception in hearing-impaired individu- als may be a particularly difficult problem for hearing aids to address. Amplification of higher harmonics may distort the speech signal, requiring the auditory systems of hear- ing-impaired people to slow down in their interpretation of human speech signals. Future hearing aids need to be designed with this problem in mind. REVIEW QUESTIONS 1. What are the important parts of the vocal tract? How are consonants and vowels produced in our vocal tract? 2. What is meant by the terms place of articulation, manner of articulation, and voicing? 3. What is coarticulation? How does its perception affect our understanding of speech? 4. What is categorical perception? How does it affect our perception of speech? What is its likely function? 5. What is the McGurk effect? How does it show the relation between vision and audition with respect to speech perception? 6. What is phonemic restoration? How does it demon- strate top-down processing in speech perception? 7. What is the difference between general-mechanism theories of speech perception and special-mecha- nism theories of speech perception? Which theory offers a better explanation of speech perception? What is the motor theory of speech perception? 8. What is perceptual narrowing? How does it help infants improve their speech perception? What problems can develop from it? 9. What areas of the brain are associated with speech production and language comprehension? What other areas of the brain may be involved in speech perception? 10. Why is speech perception difficult for those who wear hearing aids? What aspect of hearing-aid tech- nology needs to be improved to help those with mild to medium hearing impairments? PONDER FURTHER 1. Outline the several ways that the speech stimulus is ambiguous. Given that the speech stimulus is ambig- uous, discuss how perceptual mechanisms work to make speech perception work so well. 2. Examine how speech perception takes advantage of other features of auditory perception and what fea- tures of speech perception might be unique. KEY TERMS Aphasia, 359 Broca’s aphasia, 359 Broca’s area, 359 Categorical perception, 349 Coarticulation, 348 Consonants, 343 Formants, 343 General-mechanism theories, 355 International Phonetic Alphabet, 346 Larynx (voice box), 343 Manner of articulation, 344 McGurk effect, 351 Sensation and Perception368 Motor theory of speech perception, 355 Perceptual narrowing, 357 Pharynx, 343 Phonemes, 345 Phonemic restoration effect, 353 Place of articulation, 344 Special-mechanism theories, 355 Trachea (windpipe), 343 Unvoiced consonant, 344 Uvula, 343 Voice area, 361 Voiced consonant, 344 Voicing, 344 Voicing-onset time, 349 Vowels, 343 Wernicke’s aphasia, 360 Wernicke’s area, 359 Word segmentation, 353 Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review, study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning. Learning Objectives Digital Resources 12.1 Discuss the complex process of speech perception and speech production with regard to the speech stimulus. Speech Perception: Cognitive Foundations and Cortical Implementation Auditory Demonstrations Study Illuminates How Babies Learn to Speak 12.2 Sketch the mechanisms our speech perception system uses to extract coherent speech from the speech signal. Speech Perception as a Multimodal Phenomenon The Time Course of Interpretation in Speech Comprehension Sense of Touch Can Help Hearing, Study Says What’s He Saying? “Bahh” or “Fahh”? A Brain Mystery 12.3 Contrast the different types of theories of speech perception, noting their similarities and differences. How Does the Brain Decode Speech? 12.4 Diagram the areas of the brain involved in speech perception and what happens when they are damaged. Wernicke’s Aphasia Broca’s Aphasia Gusto/Science Source 13Music Perception florin1961 (Florin Cnejevici)/Alamy Stock Photo LEARNING OBJECTIVES 13.1 Explain how frequency is related to pitch, chroma, and the octave. 13.2 Summarize the basic neuroscience of music, including how training and experience can affect the representation of music in the brain. 13.3 Discuss how learning and culture affect music perception. INTRODUCTION Wherever you travel, you will find music. It may sound very different from the music you are accustomed to hearing, but you will recognize it instantly as music. In Kurdistan, we find a unique culture of music featuring such instruments as the tanbur (a fretted string instrument), the qernête (a double-reed wind instrument), and the şimşal (a flutelike instrument) (Figure 13.1). Although most of you may never have heard of these instru- ments and may never have heard Kurdish music before, you would instantly recognize them as musical instruments, and you might even like Kurdish music (see ISLE 13.1 for an example of Kurdish music). ISLE EXERCISES 13.1 Kurdish Music Example 13.2 Javanese Gamelan Music Example 13.3 Ancient Greek Music 13.4 35,000-Year-Old Flute 13.5 Is This Music? 13.6 The Octave and Tone Similarity 13.7 Pentatonic Music 13.8 Meter and Beat 13.9 Bolero Clip 13.10 Attack and Decay 13.11 Examples of Melody 13.12 Types of Scales 13.13 Gestalt Principles Review 13.14 Gestalt Principle: Proximity: Bach’s Partita No. 3 in E major 13.15 A Shave and a Haircut 13.16 Cross-Modal Matchings as a Simulation of Synesthesia 13.17. Indian Raga Music Example 13.18 Bach’s Violin Partita No. 2 in D minor 13.19 Shepard Tones 13.20 Octave Illusion 13.21 Scale Illusion 13.22 Tritone Paradox FIGURE 13.1 A Tanbur Man playing a tanbur, a traditional Kurdish instrument. © A urora Photos/A lam y 372 Sensation and Perception Better known in the United States, though further from our own musical tradition, is Javanese gamelan. Gamelan music uses a different scale system from our own, so it sounds very different. The slendro scale is a pentatonic (five-note) scale using intervals not used in Western music. Despite this difference, we certainly recognize gamelan as music (listen to ISLE 13.2 for examples of Javanese gamelan music). Figure 13.2 shows the xylophone-like instruments that are used in gamelan music. Within our own culture, we make the most of differences in musical tradition. Many of you may debate the relative merits of East Coast hip-hop and snap music, or funk metal versus nu metal (Figure 13.3). There is, however, continuity in our Western music tradition. All of these varieties of music, from classical to jazz to country to gangsta rap, use the basic Western music system, even if the performers violate Western norms of dress and body art. In this manner, most variants of rock music have more in common with the orchestral music tradition than fans of either style might care to admit. For example, both Mozart and gangsta rap use the same basic Western scale system, playing the same notes, and in mostly the same keys. One can contrast either style of music with the aforementioned gamelan music, which uses a completely different scale system. Music also has a long history. Wherever we find written records of past civilizations, we find descriptions of musical events. For example, the ancient Greeks left numerous written descriptions of music, as well as many illustra- tions of musical instruments. The story of Orpheus is as poignant today as it was then, because we can relate to the longing power of music, just as the Greeks did when the story was new. In the story, Orpheus uses music to win his love back from the dead, only to lose her again. Greeks also used a form of written music. It consisted of using certain letters to represent pitch. Although precious few examples of ancient Greek music in written form remain today, there are a few preserved parchments that allow us to re-create music from almost 2,000 years ago. The longest single written musical piece from this time period is the Seikilos song, written in approximately 200 CE. It was prob- ably not a “hit” at the time, but it is the longest segment of ancient Greek music that can be played now, thanks to its preservation in written form. (See Cartwright, 2013, or ISLE 13.3.) Traveling millennia further back in time, there are no written records of music notation. But there is a fine archaeological record that demonstrates musical instruments being made far back into the Stone Age. Wind, percussion, and string instruments all date back this far into antiquity. The oldest known instruments are flutes made from the bones of birds, some of which date back as far as 35,000 years (Conard, Malina, & Münzel, 2009). Figure 13.4 shows such flutes. We can wonder what a prehistoric man or woman might have played on one of these flutes, looking across a primordial forest landscape from the mouth of a cave, but we will never know. However, reconstructions of these flutes show that notes representing octave equiva- lence were present. You can hear a reconstruction of such a flute played in ISLE 13.4. This suggests that music has been a part of human culture for as long as humans have had culture. We can also ask this question: What is music? A dictionary might define music as an art form based on sound. But how do we know what sounds are art and what sounds are, well, just sounds? On one hand, when we hear Rihanna singing “Love on the Brain” or the Cleveland Orchestra playing Schubert’s Symphony No. 2, there is no disagreement—we are hearing music. And when we hear the sound of a washing machine whirring or the sound of landscapers mowing lawns, we know that such sounds are not music. However, ISLE 13.1 Kurdish Music Example ISLE 13.2 Javanese Gamelan Music Example © iS to ck ph ot o. co m /M ith ril FIGURE 13.2 Different Musical Cultures Gamelan musical instruments. © iS to ck ph ot o. co m /c re o7 7 FIGURE 13.3 Popular Music Rock music is popular throughout the world. ISLE 13.3 Ancient Greek Music 373 Chapter 13: Music Perception some artists stretch the limits of music. For example, in John Cage’s famous piece 4'33″, audiences “listen” to a performer doing absolutely nothing for 4 minutes and 33 seconds. The music is the rustle of people in their seats and the occasional embarrassed cough. Is this music? That depends on your perspective. Certainly, Cage wanted us to think of music in a whole new way. And what about the following? The Melbourne Symphony Orchestra played a piece of music in which every member of the orchestra was playing beer bottles instead of his or her normal musical instrument. Many of you may have also heard the typewriter symphony, which went viral on YouTube in 2012. You can see both of these pieces performed in ISLE 13.5. Are these pieces satire, or are they actu- ally music? Finally, we can also consider whether natural sounds are music. We may find many natural sounds beautiful, from birds singing to waves lapping to the wind whistling. But is birdsong music? What about the sounds of waves lap- ping on the sand on a peaceful beach? Most of us would hesitate to classify this as music, even though the sounds may be decidedly pleasing to listen to. Thus, it may actually be difficult to come up with a definition of music that satisfies all of its boundary conditions. But we will try to define music as follows: Music is ordered sound made and perceived by human beings, created in meaningful patterns (Tan, Pfordresher, & Harre, 2010). The last introductory question concerns the function of music. Why do we do it? Many scholars from many different disciplines have sought to determine the function of music for human beings, given the universality of music to the human species. Some have argued that this implies that music evolved, in the biological sense of the word, and therefore must have a function. What function might music serve in evolution? Some have argued that it served a sexual selection function. A highly musical person was likely fit, with musical ability signaling both good health and intel- ligence. In this view, music is something of a peacock’s tail: Only a fit bird can afford such an extravagant tail. By analogy, only a fit person has the time to develop musical talent. Another evolutionary view is that music serves to bind people together in coherent groups. Singing and dancing brought people together and helped them find common purpose. Of course, it is also possible that music served no evolutionary function and is just a happy by-product of the evolution of auditory processing in humans in general. The answers to these questions are beyond the scope of this book. We now turn to issues in sensation and perception. THE ACOUSTICS OF MUSIC 13.1 Explain how frequency is related to pitch, chroma, and the octave. Pitch, Chroma, and the Octave In Chapter 10, we described the relation between pitch and frequency. As frequency increases, we hear sounds at higher and higher pitches. As frequency decreases, we hear sounds at lower and lower pitches. Human beings can hear sounds between 20 and 20,000 Hz, but the range that is used in music is more restricted. Indeed, music is generally played in the range of only up to a maximum of about 5,000 Hz. A piano, for example, has its low- est note tuned at 27.5 Hz and its highest note tuned at 4,186 Hz (Figure 13.5). No Western instruments play lower than 27.5 Hz, and only a few Western instruments play higher than the piano does (e.g., the piccolo). In terms of the range at which humans can sing, the range is even more restricted. A human bass voice may get as low as 75 Hz, and a human soprano may get as high as 1,300 Hz. Thus, the notes used in music fall within a subrange of the full range of human hearing. ISLE 13.4 35,000-Year-Old Flute FIGURE 13.4 Paleolithic Musical Instruments (Conard et al., 2009) Bone flutes dating to at least 30,000 years ago. ISLE 13.5 Is This Music? Music: ordered sound made and perceived by human beings, created in meaningful patterns 374 Sensation and Perception Although the fundamental frequency of musical notes seldom exceeds 5,000 Hz, harmonics typically range higher than this. Therefore, higher frequencies are important for music perception. Remember that natural sounds, including those of all voices and musical instruments, have higher harmonics. That means that in addition to the pitch we hear, there are other sounds present at higher frequencies. These higher harmonics contribute to the experience of timbre. Thus, having recording equipment that records at higher frequencies will preserve the timbre (to be reviewed shortly) of voices and instruments. Even then, we seldom use the very high frequen- cies for music—a good thing, as most of us lose our hearing above 10,000 Hz anyway by the time we are in our 30s or 40s. The Western orchestra uses a range of instruments, some of which are designed to play lower notes and others designed to play higher notes. Among the string instruments, double basses play at the lower end of the musical range of pitches, whereas violins reach to some of the highest pitches used in music. Thus, in a string orchestra, without wind or brass instruments, the violins will play the higher notes, the violas and cellos will play intermedi- ate pitches, and the double basses will play the lower notes. Among the brass instruments, tubas and trombones play at lower pitches, whereas trumpets play at higher pitches. Among woodwinds, bassoons play at the lower end of the pitch range, whereas the piccolo has the highest notes of any instrument used in Western music. Figure 13.6 shows the frequency ranges of many common Western musical instruments. 36 9. 99 41 5. 30 34 .6 48 38 .8 91 46 .2 69 51 .9 13 58 .2 70 69 .2 69 77 .7 82 92 .4 99 10 3. 83 11 6. 54 13 8. 59 15 5. 56 18 5. 00 20 7. 65 23 3. 08 27 7. 18 31 1. 13 46 6. 16 55 4. 37 62 2. 25 73 9. 99 83 0. 61 93 2. 33 11 08 .7 12 44 .5 14 80 .1 16 61 .2 18 64 .7 22 17 .5 24 89 .0 29 60 .0 33 22 .4 37 29 .3 29 .1 35 27 .5 00 A 0 Fr eq ue nc y (H z) N ot e B 0 30 .8 63 C1 32 .7 03 D 1 36 .7 08 E1 41 .2 03 F1 43 .6 54 G 1 48 .9 99 A 1 55 .0 00 B 1 61 .7 35 C2 65 .4 06 D 2 73 .4 16 E2 82 .4 07 F2 87 .3 07 G 2 97 .9 99 A 2 11 0. 00 B 2 12 3. 47 C3 13 0. 81 D 3 14 6. 83 E3 16 4. 81 F3 17 4. 61 G 3 19 6. 00 A 3 22 0. 00 B 3 24 6. 94 C4 26 1. 63 D 4 29 3. 67 E4 32 9. 63 F4 34 9. 23 G 4 39 2. 00 A 4 44 0. 00 B 4 49 3. 88 C5 52 3. 25 D 5 58 7. 33 E5 65 9. 26 F5 69 8. 46 G 5 78 3. 99 A 5 88 0. 00 B 5 98 7. 77 C6 10 46 .5 D 6 11 74 .7 E6 13 18 .5 F6 13 96 .9 G 6 15 68 .0 A 6 17 60 .0 B 6 19 75 .5 C7 20 93 .0 D 7 23 49 .3 E7 26 37 .0 F7 27 93 .0 G 7 31 36 .0 A 7 35 20 .0 B 7 39 51 .1 C8 41 86 .0 FIGURE 13.5 Piano Keyboard A piano’s notes start at 27.5 Hz at the far left and go up to 4,186 Hz at the far right. Most music is played within this range. A 27 .5 00 B 30 .8 63 C 32 .7 03 D 36 .7 08 E 41 .2 03 F 43 .6 54 G 48 .9 99 A 55 .0 00 B 61 .7 35 C 65 .4 06 D 73 .4 16 E 82 .4 07 F 87 .3 07 G 97 .9 99 A 11 0. 00 B 12 3. 47 C 13 0. 81 D 14 6. 83 E 16 4. 81 F 17 4. 61 G 19 6. 00 A 22 0. 00 B 24 6. 94 C 26 1. 63 M id dl e C D 29 3. 66 E 32 9. 63 F 34 9. 23 G 39 2. 00 A 44 0. 00 A –4 40 B 49 3. 88 C 52 3. 25 D 58 7. 33 E 65 9. 26 F 69 8. 46 G 78 3. 99 A 88 0. 00 B 98 7. 77 C 10 46 .5 D 11 74 .7 E 13 18 .5 F 13 96 .9 G 15 68 .0 A 17 60 .0 B 19 75 .5 C 20 93 .0 D 23 49 .3 E 26 37 .0 F 27 93 .0 G 31 36 .0 A 35 20 .0 B 39 51 .1 C 41 86 .0 29 .1 35 34 .6 48 38 .8 91 69 .2 69 77 .7 82 13 8. 59 15 5. 56 27 7. 18 31 1. 13 55 4. 37 62 2. 25 11 08 .7 12 44 .5 22 17 .5 24 89 .0 46 .2 49 51 .9 13 58 .2 70 92 .4 99 10 3. 83 11 6. 54 18 5. 00 20 7. 65 23 3. 08 36 9. 99 41 5. 30 46 6. 16 73 9. 99 83 0. 61 93 2. 33 14 80 .0 16 61 .2 18 64 .7 29 60 .0 33 22 .4 37 29 .3 Timpani Trumpet Flute Chimes Guitar Bass FIGURE 13.6 The Symphony Orchestra: Frequency and Pitch This diagram shows several instruments and their frequency ranges. 375 Chapter 13: Music Perception The Octave Pitches get higher as frequencies go up, but there is another important dimension of pitch that is highly relevant to music: the octave. The octave is the interval between one note and a note with either double the frequency or half the frequency of that note. This is a physical definition: A frequency of 200 Hz has octaves of 100 Hz below it and 400 Hz above it. Psychologically, we hear similarities between these doubled or halved frequencies. In musical terms, we refer to them by the same note name, but at different octaves. Thus, even though notes at 200 and 220 Hz are more similar in pitch, we hear notes at 200 and 400 Hz as being alike in a way that notes at 200 and 220 Hz are not (to hear this for yourself, go to ISLE 13.6). We hear notes that are an octave apart as similar in a fundamental way. That is, there are perceived similarities between sounds that are an octave apart from one another. Notes that are one octave apart are said to be of the same chroma. This concept of an octave is present in all Western and non-Western musical traditions, including, as we mentioned earlier, in the functioning of prehistoric musical instru- ments. When we hear two notes that are an octave apart, they sound similar to us, despite their difference in pitch. On a piano keyboard, we label notes an octave apart with the same name. Thus, middle C has a frequency of 261.6 Hz, and the C one octave above it has a frequency of 523.3 Hz, or approximately double that frequency. The next C is at 1,046.5 Hz. Any musician recognizes that a scale begins at one note and continues to the same note at the next octave. Think now about a piano keyboard and examine the illustra- tion in Figure 13.5. We see a pattern of white and black keys. You will see the white keys labeled as being one of seven chromas of notes, C, D, E, F, G, A, and B, and then the pattern repeats back to C. This pattern repeats across the piano keyboard. Each C shares a feature of sound in common with other C’s, but not with other notes. Similarly, each G shares a feature of sound in common with other G’s, but not with other notes. This feature that these notes share in common is that they sound similar to us, and we call this feature chroma. For example, middle C (262 Hz) is closer in fre- quency to the D just above it (294 Hz), but it sounds more similar to the C in the next octave (523 Hz). This similarity of chroma from one octave to the next is represented by the pitch helix shown in Figure 13.7. As one goes up the helix, pitches get progressively higher, but the twists in the helix indi- cate the octave equivalence across similar notes. Returning to the piano keyboard, we also have the black keys, which represent “sharps” and “flats” in musical terms. These keys are at frequencies between the white keys to either side. Thus, the black key between middle C and the D next to it plays at a frequency of 277 Hz, approximately halfway between the frequencies of the C and the D. In musical notation, this black key can be called either C-sharp or D-flat. The black key between the D and the E can be called either D-sharp or E-flat. Whether the note is called by its sharp name or its flat name depends on the musical context, but the sound is the same. When we add the sharps and flats to our musical hierarchy, we have 12 notes in an octave, as we ascend from one note to the same note an octave higher. Each adjacent note is sometimes called a semitone. There are 12 semitones in an octave in Western music. In music, when every note, including the sharps and flats, is played between one octave and the next (i.e., every semitone), this is called the chro- matic scale. This would mean playing a 13-note scale, starting, for example, with C and including and ending with the C one octave above it. Almost all Western instruments ISLE 13.6 The Octave and Tone Similarity Octave: the interval between one note and a note with either double the frequency or half the frequency of that note Chroma: the subjective quality of a pitch; we judge sounds an octave apart to be of the same chroma Semitones: the 12 equivalent intervals or notes within each octave increasin g fre que nc y E4F4 G4 G3 G2 G1 F3 F2 F1 A1 Aligned notes have same chroma Color indicates tone chroma A2 A3 A4 A5 E3 E2 E1 B1 C1 D1 C2 Tone heightD2 C3 D3 C4 C5 D4 B2 B3 B4 B5 FIGURE 13.7 Pitch Helix (Shepard, 1982) The pitch helix shows the relation of both pitch and frequency and of pitch similarity across octaves. 376 Sensation and Perception allow musicians to play all 12 notes of the octave. Exceptions include some kinds of harps and recorders. The other feature of the Western musical tradition is the use of an equal-temperament scale. This means that every adjacent note has an identical frequency ratio. The absolute difference between adjacent notes increases as one gets higher in frequency, but the ratio matters for per- ception. In this way, we perceive the difference between each successive semitone as equivalent in terms of difference in pitch to the one before it. This demonstrates Weber’s law. What matters in perception is the ratio, not the absolute difference. One advantage of the equal-temperament system is that any melody can be played starting on any particular note. In most Western music, the differences in frequency between each note are well established and do not vary (this may change in some music, such as a cappella choirs). When musicians tune their instruments, they tune them so that their C’s (or, usually, A above middle C) all match at the same pitch or frequency. A piano or tuning whistle often provides this frequency. This organization has become standardized across Western music. Thus, a violinist in California and a cello player in Zurich, Switzerland, will usually agree that the A above middle C is tuned to 440 Hz (as long as they are not specifically tuning to standards from centuries past, as some orchestras do). Traditional Chinese music uses a different scale system. Instead of the diatonic (eight-note) scale used in Western music (C, D, E, F, G, A, B, and C), Chinese music uses only a five-note (pentatonic) scale. In addition, the notes are not tuned according to an equal-temperament system, so that one cannot play the same melody starting on a different note, because the ratios between successive notes are not the same. Some early 20th-century clas- sical music, in trying to defy convention, essentially used pentatonic scales as well. Figure 13.8 shows a woman playing a guqin, a traditional Chinese instrument. The guqin is a seven-string instrument similar to a zither. Traditionally, it was tuned to the Chinese scale known as zheng diao, a pentatonic scale, although most instrumentalists use Western-based notes today. Westerners typically find traditional Chinese music a bit odd because the notes do not map directly onto the notes in our scale, which we have become so accustomed to hearing. Some Western forms of music use pentatonic scales, but these versions use the notes or pitches used in the equal-temperament scale system. These pentatonic tradi- tions include Celtic folk music, some forms of West African music, and the American blues tradition. The five-note tradition makes improvisation, a hallmark of both Celtic music and American blues, easier. ISLE 13.7 gives examples of both Celtic music and improvisation in the American blues tradition. To summarize, pitch is the psychological experience of frequency. As frequency gets higher, we hear the sound at a higher pitch. Musical notes are set at particular frequencies, and the relations between notes in Western music follow an equal-temperament system. TEST YOUR KNOWLEDGE 1. What is meant by the term chroma? How is chroma related to the pitch helix? 2. What is meant by the term semitone? What is an equal-temperament scale? Consonance and Dissonance In most music, more than one note is played at the same time. This is true in any musical style. Even a lonely folk singer and his or her guitar is playing at least two notes (one in voice, one on guitar) at the same time. A piano player has 10 fingers and thus can Equal-temperament scale: a tuning system in which the difference between each successive semitone is constant both in pitch and in frequency FIGURE 13.8 A Guqin A woman playing a guqin, a traditional Chinese instrument. © B lu e Je an Im ag es /A la m y ISLE 13.7 Pentatonic Music 377 Chapter 13: Music Perception play 10 notes at the same time, though this is very rare. A symphony orchestra may have many instruments playing several different notes at the same time. Jazz bands may have a number of different musical instruments playing at once. Rock bands tend to have singing, guitar playing, and percussion simultaneously. How do composers know which notes will sound good to listeners when played at the same time as other notes? The concept of harmony in music refers to which pitches sound pleasing when played together. In technical terms, consonance refers the perception of pleasantness or harmony when two or more notes are played; that is, the notes fit with each other. In con- trast, dissonance refers to the perception of unpleasantness or disharmony when two or more notes do not fit together. Why some notes are consonant when played together whereas others are dissonant has been the subject of much debate within Western cul- ture, with theories going back all the way to the time of the ancient Greeks. The Greeks were impressed that two tones that could be expressed as a simple ratio of each other tended to sound consonant, whereas those that were more complex tended to sound dissonant. They did not know about frequency, but they measured these ratios in terms of the lengths of vibrating strings. Thus, the Greeks knew that a vibrating string twice the length of another would produce a consonant octave sound. We now know that two notes separated by an octave have approximately a 2:1 ratio of frequency. For example, concert A is 440 Hz, and the A above it is 880 Hz. Similarly, intervals of a major third (e.g., C and E) and a perfect fourth (e.g., C and F) sound consonant, but adjacent notes (e.g., C and D, a major second) sound dissonant. Major thirds and perfect fourths are easy to express as ratios, whereas adjacent notes are not. When more than two notes are played at the same time, the result is called a chord. Chords are often played on the piano with the left hand, while the right hand plays a melody. Fundamental training in music allows musicians to learn which chords are con- sonant and which are dissonant. In many pieces of music, chords are selected to be har- monious or consonant with the melody line. We will discuss melody later in this section. Musical context also plays a role in our perception of consonance and dissonance. There may be some musical situations in which adjacent notes go together and would sound consonant, so consonance goes beyond simple ratios. In addition, culture plays a role in our perception of consonance and dissonance. What we find consonant in Western culture might be dissonant in traditional Chinese music, and what traditional Chinese music deems consonant we might find dissonant. In addition, norms within a culture change over time. This is true of the major third in Western music. Prior to Bach’s time, the major third was avoided, as it was considered dissonant. It is now con- sidered the most consonant interval in Western music after the octave itself. Dynamics and Rhythm Music is not just a series of pitches. Equally important in the production and appre- ciation of music are dynamics and rhythm. Indeed, drum music may not vary at all in pitch—the differences are in the complex rhythms. We start by defining the rele- vant terms in this section. Dynamics refers to the relative loudness and how loudness changes across a composition. That is, a piece may start off very loud, then grow softer, and then finish loud again. Changing from loud to soft may be important in trans- mitting the meaning and emotion in any piece of music. In musical notation, soft is indicated by a p for piano, and loud is indicated by an f for forte (these are the Italian terms for “soft” and “loud”). In physics terms, dynamics refers to amplitude, measured in decibels. Forte means more decibels, whereas piano means fewer decibels. Rhythm refers to the temporal patterning of the music, including the tempo, the meter, and the beat. Tempo refers to how fast or slow a piece of music is played, that is, the speed of the music. For example, a beginning musician may elect to play a piece at a Harmony: the pleasant sound that results when two or more notes are played together Consonance: the perception of pleasantness or harmony when two or more notes are played; that is, the notes fit with each other Dissonance: the perception of unpleasantness or disharmony when two or more notes do not fit together Dynamics: relative loudness and how loudness changes across a composition Rhythm: the temporal patterning of music, including the tempo, the beat, and the meter Tempo: the pace at which a piece of music is played 378 Sensation and Perception slower tempo so as not to make mistakes, whereas a more experienced musician may play the piece faster. Tempo can also change within a piece. Usually brisk or fast tempos are used to express joy, whereas slower tempos render a sadder feeling. Think of Christmas music. “Rudolf, the Red-Nosed Reindeer” is played quickly to express joy, whereas “Silent Night” is a slow piece that reflects a more thoughtful or religious approach to the holiday. Meter refers to the temporal pattern of sound across time, which usually repeats itself across the piece. Meter is completely intertwined with beat. Beat refers to spaced pulses that indicate if a piece is fast or slow. Thus, meter tells you how many beats occur per musical measure (the repeating temporal pattern), and beat tells you which notes to emphasize. In rock music, drums usually “keep the beat” by pulsing throughout each measure. In traditional classical music, instruments such as the double bass are respon- sible for keeping the beat. In most popular music, as well as in marches and many other styles of music, the meter is called 4/4, meaning that there are four beats per measure. In this meter, there is usually an emphasis, often indicated by relative loudness, on the first beat out of every four and a secondary emphasis on the third beat in each measure. Waltzes are played in 3/4 time, with the emphasis placed on the first beat out of every three in a measure. The characteristic feature of Jamaican reggae music is that instead of the first beat getting the emphasis in a 4/4 measure, the second and the fourth beats out of every four get the emphasis in each measure. If you hum the melody to such nearly uni- versally known tunes as Bob Marley’s “Jammin,” you can feel the pulses on those second and fourth beats (see ISLE 13.8 for some examples of meter and beat). Rhythm is therefore a complicated feature of music. It refers to tempo, meter, and beat. In any given piece of music, each note or pitch may also be maintained for either a short period of time or a long period of time. That is, a note, such as B-flat, may be played for just one beat, or it may be sustained across four or more beats. The pattern of notes across these beats also contributes to rhythm. In jazz, for example, a common motif is a tendency to have a slightly longer note followed by a slightly shorter note, typically eighth notes. These eighth notes are indicated by the same musical notation, but a jazz musician automatically plays the first note for a bit longer and shortens the second note. This pattern gives jazz its characteristic rhythm (Figure 13.9). Waltzes usu- ally contrast a note played on the first beat with notes played on the other two beats, to give waltzes their particular 3-count rhythm, which also makes them easy to dance to. Timbre Timbre refers to the complex sound created by harmonics (see Chapter 10 if this defi- nition does not make sense). For example, a violin and a flute may be playing a note with the same pitch, but it sounds different on each instrument. The harmonics, as well as attack and decay characteristics, give each voice and each instrument its own dis- tinct sound. Composers will select specific musical instruments because of their timbres. Depending on context, specific timbres of different instruments will convey particular meanings or emotions. The oboe, for example, is often used to express sadness, bitter- sweet emotion, and perhaps puzzlement, whereas a flute is more likely to express joy. In his famous piece Bolero (1928), Maurice Ravel has different instruments play the same theme repeatedly. Each instrument gives the theme a different feel, as Ravel builds up to finally having all the strings play the theme together and then the entire orchestra (you can hear this piece on ISLE 13.9). Ravel’s Bolero also neatly illustrates a number of other principles. It is written in 3/4 meter, and you can hear the emphasis on the first beat of every measure. Moreover, as the piece progresses, the dynamics change, and the piece gradually builds from very soft to very loud. Meter: the temporal pattern of sound across time Beat: spaced pulses that indicate if a piece is fast or slow ISLE 13.8 Meter and Beat FIGURE 13.9 Louis Armstrong Louis Armstrong (1901–1971) was a famous trumpet player and singer in the jazz tradition. Jazz has roots in Western art music, popular music, and West African musical traditions. © H an k M or ga n/ Sc ie nc e So ur ce . ISLE 13.9 Bolero Clip 379 Chapter 13: Music Perception As stated earlier, the fundamen- tal frequencies of music predomi- nantly fall below 5,000 Hz. Only the piccolo and piano even come close to that frequency. However, the harmonics of musical notes often exceed 5,000 Hz, and these har- monics contribute to the timbre of a voice or instrument. For this reason, recording equipment should be able to record frequencies well in excess of 5,000 Hz in order to capture the full complexity and musicality of any musical piece, even though we do not perceive these high-frequency harmonics as actual pitches (Figure 13.10). Timbre is also important in distinguishing between well-made and poorly made instruments. The materials and craftsmanship that go into a well-made instrument allow harmonics to be created, each at the right level of loudness. Thus, well-made instruments sound better than poorly made ones, assuming the musician playing each one is of equal ability. That is, the same good musician playing on a fine violin relative to a cheap violin will sound much better on the fine violin. The well-made violin has a rich and deep timbre, even when high notes are being played, whereas the cheap violin will sound shrill on higher notes, even when played by an expert. It is for this reason that violinists favor well-made violins, including such famous antique violins as the legendary Stradivarius violins. Because of differences in timbre from instrument to instrument, the price differences between well-made and poorly made instruments can be shocking. A beginner’s violin might cost as little as U.S. $80, whereas violins made for professionals usually run higher than U.S. $30,000 (and even higher; some violins cost millions). Similarly, a student’s saxophone may cost as little as U.S. $200, whereas a professional one may cost more than U.S. $8,000. It is also clear that harmonics are not the only factor that affects timbre. Two other important features of timbre are attack and decay. Attack refers to the beginning buildup of a note. This means how quickly the instrument expresses all of its frequen- cies and if there are any differences in the onset of harmonics. Decay refers to how long the fundamental frequency and harmonics remain at their peak loudness until they start to disappear. For example, a trumpet has a very fast attack, leading to the sharp sound we associate with trumpets. In electronic instruments, attack and decay can be altered to mimic the sounds of other instruments or to create timbres that are not possible using string or wind instruments. You can hear differences in attack and decay by going to ISLE 13.10. That gives us the basic building blocks of music—pitch, loudness, rhythm, and tim- bre. Musicians combine these building blocks in infinite ways to create music of all kinds. But to understand the perception of music requires more than a description of the building blocks—it also requires a more gestalt approach, as music transpires over time. For this reason, melody is of the utmost importance. We turn to melody in the next section. TEST YOUR KNOWLEDGE 1. What is the difference between rhythm, tempo, and beat? 2. What is timbre? What physical differences produce differences in timbre? FIGURE 13.10 Sound spectrograms of two different instruments, an oboe (a) and a flute (b), playing the same note. A m p li tu d e ( d B ) Frequency (Hz) 100 30 20 10 0 1000 10,000 a. Note with frequency of 1046 Hz, played on the oboe A m p li tu d e ( d B ) Frequency (Hz) 100 30 20 10 0 1000 10,000 b. Note with frequency of 1046 Hz, played on the �ute Attack: the beginning buildup of a note Decay: how long the fundamental frequency and harmonics remain at their peak loudness until they start to disappear ISLE 13.10 Attack and Decay 380 Sensation and Perception Melody Most people can hum a variety of melodies, from tunes learned in childhood, such as “Mary Had a Little Lamb,” to Christmas songs to famous classical melodies, such as Beethoven’s Ode to Joy, to the melody of the current hot songs on the Top 40. If you think about these tunes, you may realize that melody is essentially a series of pitches joined together with rhythm created by different lengths of each note (Figure 13.11). Thus, we can define melody as a rhythmically organized sequence of notes, which we perceive as a single musical unit or idea. What carries melody beyond pitch and rhythm is that the sequence forms a unit with properties that transcend the individ- ual pitches and lengths of notes. A melody coheres in time to create an experience in its listeners. Thus, the two melodies in Figure 13.11 are very similar in terms of the notes used and the rhythms used, but anyone brought up in Western culture would never confuse these two melodies (also see ISLE 13.11). Most music, in any tradition, starts with a melody, usually sung by a voice, played on the piano or with other instruments, which is then augmented by various musical accompaniments. Untrained listeners focus first on the melody. Scales and Keys and Their Relation to Melody Consider the piano keyboard again (see Figure 13.5). If we start on middle C and play every white note to the next C, we have played a C major scale. A scale is a set of ordered notes start- ing at one note and ending at the same note one octave higher. In this way, a scale is a very simple melody. In Western music, major scales refer to sequences of notes with the following pattern of semitones: 2, 2, 1, 2, 2, 2, 1. The numeral 2 means that we go up two semitones, whereas the numeral 1 means that we go up one semitone. Thus, a G major scale starting on G will include one black note (F-sharp). One can start a major scale on any note on the piano and follow this sequence. For example, the C-sharp major scale will have the following notes: C-sharp, D-sharp, F (E-sharp), F-sharp, G-sharp, A-sharp, C (B-sharp), and C-sharp. In essence, you can start a major scale on any note if you follow the pattern of 2, 2, 1, 2, 2, 2, 1 on your piano keyboard or any other instrument. Major scales are among the first melodies any instrumentalist learns when first starting to learn to play an instrument, and the major scale is the most common kind of scale in Western music. (You can hear an assortment of scales in ISLE 13.12.) Major scales can be contrasted with the chromatic scale, in which every step is one semitone. Thus, the chromatic scale cannot be divided into keys, because it does not matter where you start or stop—the sequence is always the same. Each note is one semitone higher or lower than the previous one. On a piano, a chromatic scale means playing every key, including all of the black keys (again, see ISLE 13.12). There are also a number of different types of minor scales, which have different sequences of semitones as one moves from one octave to the next. For example, the nat- ural minor scale has the following sequence: 2, 1, 2, 2, 1, 2, 2. The natural minor scale is relatively simple in both its sequence and its relation to the major scale. However, more commonly used in music is the harmonic minor scale, with the following sequence: 2, 1, 2, 2, 1, 3, 1. The harmonic minor scale is often used in Western music to express sad- ness. It is also commonly used in Middle Eastern music. Other minor scales exist, but they are seldom used in music today. If you have taken a course in music theory, how- ever, you are familiar with the whole family of minor scales. (a) (b) FIGURE 13.11 Musical Notation (a) The musical notation for “Mary Had a Little Lamb.” (b) The musical notation for Beethoven’s Ode to Joy. These pieces are seldom grouped together because of their different histories, but they are quite similar in terms of their melodies. ISLE 13.11 Examples of Melody ISLE 13.12 Types of Scales Melody: a rhythmically organized sequence of notes, which we perceive as a single musical unit or idea Scale: a set of ordered notes starting at one note and ending at the same note one octave higher 381 Chapter 13: Music Perception Any particular melody can be expressed in terms of its key signature, which relates the melody to the pattern of scales described in the previous paragraphs. That is, every melody is played in a particular key, which refers to the main scale pattern. If we are playing “Mary Had a Little Lamb” in the key of C major, this means that the tonic is C, and we are not likely to have any sharps or flats (i.e., no black keys on the piano). Thus, key refers to the tonic note (e.g., C in a C major or minor scale) that gives a subjective sense of arrival and rest in a musical piece. Because melodies are defined in terms of the pattern of notes relative to other notes, any melody can be played in any key. Thus, “Mary Had a Little Lamb” can be played in the key of C, the key of G, or any other key. If you hum the piece, its last note is the tonic of that key. If you are singing it in the key of C major, the tonic will be C. If you switch to another key, the name of that key will be the tonic. Composers will make deviations from the key. Thus, when a G-sharp is called for in a melody played in the key of C, the G-sharp is called an “accidental” by musicians. Thus, for musicians, the term accidental refers to a note that requires a special mark to remind the musician to play the sharp or flat not present in that key. In most cases, what defines a melody is the relation of pitches within a piece rather than the absolute pitches. For example, it really does not matter what note we start “Mary Had a Little Lamb” on, as long as the remaining notes show the same relation to that note as in the original version. For example, “Mary Had a Little Lamb” is shown in the key of G in Figure 13.11. In this key, the first note is B, the second is A, and the third is G. Each of these notes is one step (or interval or two semitones) higher than the next note in the sequence. If we switch the key to F, the first three notes will be A, G, and F. Although there is no overlap in actual notes, we hear these sequences as being the same, and both can create the melody of “Mary Had a Little Lamb.” That there can be two or more versions of a melody, each starting on a different note, is known as a transposition in music. Trained musicians may be able to detect what key a simple melody is being played in, but most listeners hear the melody as such but do not register the key or starting note. Our perception of melody across transpositions starts very early in life. Plantinga and Trainor (2005) examined melody perception in 6-month-old infants. In the study, Plantinga and Trainor played particular melodies to infants numerous times over a 7-day period. On the next day, the infants either heard the same melody, but transposed into another key, or a novel melody. If the transposed melody was heard as the same melody as the original, the infant would not consider it novel and would look toward the source of the novel melody. If, however, the infant heard the transposed melody as novel, the infant would show no difference in looking time toward the source of the transposed melody or the new melody. The results showed that the infants looked more often toward the source of the novel melody, rather than the transposed melody, indicating that they perceived the transposed melody as being similar to something heard earlier. In this way, we can assert that even young infants hear melodies across transpositions of key. Gestalt Principles of Melody Because of the importance to melody of the relation among notes rather than absolute pitch, and because the perception of melody is qualitatively different than the percep- tion of a string of pitches, melody perception lends itself to the use of gestalt principles. If you remember from Chapter 5, gestalt psychology approaches perception by exam- ining emergent properties that can be seen across a perceptual array but may not be obvious in any particular single element of that stimulus. That is, the motto of gestalt psychology is that the “whole is bigger than the sum of the parts.” The gestalt principles described in Chapter 5 are certainly applicable to melody perception (Tan et al., 2010). We review these principles and then apply them to music (also see ISLE 13.13). Key: the tonic note (e.g., C in a C major or minor scale) that gives a subjective sense of arrival and rest in a musical piece Transposition: the process through which one can create multiple versions of a melody that start on different notes but contain the same intervals or sequence of changes in notes 382 Sensation and Perception The four principles are as follows: 1. Proximity: Elements near each other are seen as a group. 2. Similarity: Elements that are similar are seen as a group. 3. Closure: An incomplete pattern is seen as whole when the completion occurs. 4. Good continuation: Smooth continuity is preferred over changes in direction. These principles are applied in melody processing over time rather than space (e.g., visual processing). Indeed, music transpires over time—a melody is a sequence of notes in time. For example, we tend to hear notes that are close in chroma as grouped together, even when they come from different instruments or different locations, a source of the music illusions described at the end of the chapter. Of course, other aspects of music involve space—the source of a voice or instrumental sound is often critical—but the time dimension comes first. Space does play a role in music perception, as anyone who has heard an old monophonic record can attest. The old monophonic recordings lack the depth provided by modern recordings, which allow listeners to imagine where different instruments’ sounds are coming from. But returning to the gestalt principles, we start with proximity. We then consider similarity and closure. Proximity. In music, proximity may refer to elements’ being close together in pitch, time, or space (Tan et al., 2010). For example, notes that are similar in pitch may be grouped together. Notes are also grouped together if they are played together in time or if they come from the same instrument or section of a larger musical group. To get a sense for the idea of proximity, imagine a person playing a piano. Typically, the right hand plays notes that are higher in pitch than the left hand. Also, most often, it is the right hand that plays the melody, whereas the left hand plays the bass line or accompaniment. Even though all of the notes are played in close spatial proximity and at approximately the same time, we hear the notes from the right hand emerging as melody because they are grouped together with respect to pitch (Figure 13.12). Similarly, in some of Bach’s famous solo music for violin, the violinist essentially creates two streams of music by simultaneously playing both high and lower notes. Perceptually, we group the high notes together and group the low notes together, so we hear it as polyphonic or as two lines of music (for an example, go to ISLE 13.14 to hear Bach’s Partita No. 3 in E major). Similarity. Think of listening to your favorite music. Chances are your favorite music is created by a group of musicians, playing different parts. From hometown garage bands to the Vienna Philharmonic, most music consists of mul- tiple parts. A rock band may have a drummer, a guitar player, a bass player, and one or more singers (Figure 13.13). An orchestra may consist of more than 100 musicians playing 16 different instruments. Composers will use our perception of similarity to create seamless perceptions of melody even when the melody crosses from one voice or one instrument to another. Similarity plays out at several levels. We may hear similar timbres grouped together. A modern orchestra may have 16 vio- linists all playing the same part. Because these musicians are playing the same notes with the same approximate timbre, we hear them as grouped together. Moreover, once a melody has been established, we may follow the melody because of its similarity across changes in instruments. In Ravel’s Bolero, the instruments playing the melody constantly change, but we have no difficulty distinguishing the melody from the bass line, drum rhythms, and harmony (hear this in ISLE 13.9). ISLE 13.13 Gestalt Principles Review C D C B FIGURE 13.12 Notes Grouped Together by Proximity Because C and D are close in pitch, they are grouped together. Because C and B in a higher octave are not close in pitch, we think of them as separate. ISLE 13.14 Gestalt Principle: Proximity: Bach’s Partita No. 3 in E major © iS to ck ph ot o. co m /o ne in ch pu nc h FIGURE 13.13 A Rock Band A rock band may have a drummer, a guitar player, a bass player, and one or more singers. We infer notes that are similar in pitch as coming from the same musician. 383 Chapter 13: Music Perception Closure. In music, closure means that a melody should end on the tonic note of any particular scale or another note implied by the progression of the melody. Typically, if the melody is played in the key of C, the last note will be C. Occasionally, the note might be G, but seldom any other note in the key of C. To illustrate this point, think of the very short melody of the song “Shave and a Haircut.” If you do not know the melody, but can read music, see Figure 13.14 (you can hear it in ISLE 13.15). If you simply play the notes or sing the words “shave and a haircut,” most people experience a strong longing to hear or sing the last two notes (“two bits”). Try it yourself: Most people cannot stop themselves from singing the last two notes of the sequence. In the 1988 movie Who Framed Roger Rabbit? the character of Roger Rabbit is lured out of hiding because he cannot stop himself from completing this melody. The importance of closure in melody perception was demonstrated in an experiment by DeWitt and Samuel (1990). Their experiment was a musical demonstration of the phonemic restoration effect we discussed in the last chapter, except instead of hearing a missing phoneme, participants heard an implied but missing note. In this perceptual restoration effect, participants heard a major scale plus the next two notes, for a total of 10 notes, each predictable from the note played before it. One note of the 10 was replaced by white noise. Under these conditions, many participants reported hearing the missing note. This effect was greater when the note was later in the scale, allowing more time for expectations to build up. The effect did not occur if the notes were ran- domly arranged rather than in a melody-like scale. Thus, our expectation of what to hear in a particular sequence can actually create the perception of that note. Good continuation. Most composers will tend to have one note be relatively close to the previous note in pitch. That is, a C is more likely to be followed by a D than by an F. This allows the listener to hear those notes as connected. Some composers (e.g., Bach) will often create two lines within a musical piece by alternating between a high note and a low note successively. Because of continuation, we hear the lower notes as one line and the higher notes as another. TEST YOUR KNOWLEDGE 1. Why are melodies easy to transpose in the Western system of music-making? 2. How do each of the gestalt principles apply to the concept of melody? THE NEUROSCIENCE OF MUSIC 13.2 Summarize the basic neuroscience of music, including how training and experience can affect the representation of music in the brain. One of the central tenets of modern psychology is that perceptual and cognitive pro- cesses arise in the brain. Even if these processes are about objects in the world and per- ceived through sensory organs, the brain’s role is still considered central. Implicit in this view is that all human brains are organized in similar ways. This turns out to be sup- ported over and over again. For example, regardless of gender, age, ethnic background, racial identification, native language, and any other feature that distinguishes one human from another, brain regions serve the same functions across these distinctions. For example, the occipital lobe processes visual information for all sighted persons Shave and a hair-cut two bits FIGURE 13.14 Principle of Closure It is hard not to want closure of this famous, albeit very short, piece of music. ISLE 13.15 A Shave and a Haircut 384 Sensation and Perception (indeed, all sighted mammals). Wernicke’s area and Broca’s area are language areas in all human brains. The primary auditory cortex processes input from the cochlea in all human brains. Neurosurgeons may need to be aware of slight deviations in individual human brains—one region may be shifted a few millimeters forward or backward rel- ative to other individuals—but by and large, this rule of brain universalism has been upheld in modern science. The Neuroanatomy of Music The neuroanatomy of music is a bit different because music is in some ways an optional feature of human cognition. Although all cultures have music (including those that for- bid it), people vary greatly in their interest in music, the time spent listening to music, their individual training in music, and the musical traditions they are exposed to. As a result, there are greater individual differences in the regions responsible for music per- ception and in the way these regions function than there are for many other functions of the brain. Nonetheless, we can make some generalizations about how music perception occurs in the brain. First, we start with a quick review of information covered in Chapters 10 and 11. Recall that from the cochlea to the auditory cortex, the auditory signal has a tonotopic organization. This means that the auditory nerve preserves a representation of the fre- quency of sound, which we perceive in music as the pitch of a particular note. Loudness is represented by the strength of the signal at any particular frequency. We also see tonotopic organization throughout the primary auditory cortex. Music perception is certainly a variant of sound perception, but not all sounds are music. Thus, we can ask if there are common areas of the brain that process musical stimuli, independent of other sounds, in all human beings. We find these areas starting in the sec- ondary auditory cortex in the temporal lobe (Overy, Peretz, Zatorre, Lopez, & Majno, 2012). One of the clear findings is that music perception usually causes greater activation in the right temporal lobe than the left tempo- ral lobe (Overy et al., 2012). It seems that the right hemisphere is more sensitive to small changes in pitch, which are likely relevant to music but less relevant to speech. For example, Hyde, Peretz, and Zatorre (2008) used functional magnetic resonance imaging (fMRI) to examine the function of the right and left auditory cortical regions in frequency processing of melodic sequences. They found that better behavioral pitch resolution was associated with activity in the right secondary auditory cortex. More specifically, these areas included the planum temporale as well as some areas within the primary auditory cortex (Figures 13.15 and 13.16). Indeed, many neuroimaging studies have now shown the importance of the right secondary auditory cortex in pitch per- ception in music (Janata et al., 2002; Overy et al., 2012). However, the more musical training an individual has, the more the left hemisphere is involved in music perception (Habibi, Wirantana, & Starr, 2013). This is true in children as well as adults (Habibi, Cahn, Damasio, & Damasio, 2016). That is, musical training forces the brain to devote more networks to music and perhaps requires the brain to involve meaning and lan- guage circuits to help with music perception and cognition. Indeed, many argue that musical training affects the way we hear music as well. Other neuroimaging studies have focused on harmonic expectations. For exam- ple, Seger et al. (2013) played small samples of Western classical music to participants while they were being monitored by fMRI. Compared with when participants were not FIGURE 13.15 Secondary Auditory Cortex Secondary auditory cortex areas with specialization in fine pitch perception are important in music perception. Secondary auditory cortex Primary auditory cortex 385 Chapter 13: Music Perception listening to music, Seger et al. found activity in the bilateral superior temporal gyrus and the right inferior frontal gyrus. These areas also became even more active when the experimenters changed the music to violate expected patterns. Another important component of music, as previously discussed, is rhythm. Rhythm appears to be processed in areas of the primary auditory cortex and, more noticeably, in the right hemisphere as well. In particular, the belt and parabelt areas are important in the process- ing of rhythm (Snyder & Large, 2005; Tramo, 2001). Moreover, when people are producing rhythm, we also see some more prominent activ- ity in the left hemisphere, including areas of both the left prefrontal cortex and the left parietal cortex. Because people producing rhythm are also engaged in action, we also see activity in the cerebellum (Tramo, 2001). Varieties of musical training, ranging from learning to play the violin to learning to play the drums, may also impact neural development (Slater, Azem, Nicol, Swedenborg, & Kraus, 2017). Most of the studies reviewed in this section were conducted with participants who were college students without specific musical train- ing. Most college students have spent many hours over the course of their lives listening and attending to music. But we can also ask how musical training affects the networks in the brain for music perception. Although there are a great many similarities between the brains of nonmusicians and musicians, we can find some important differences as a function of musical training. For example, the organization of the motor cortex changes in response to the demands of the complex motor movements needed to play many instruments. In particular, Krings et al. (2000) examined the brain areas used by professional piano players and a control group. Piano players must use com- plex movements of both hands, and in many cases, the movements of each hand may follow very different patterns simultaneously. Using fMRI, they found that the professional piano players required lower levels of cortical activation in motor areas of the brain relative to controls while doing the same task. That is, musical training allowed greater control of the hands in piano players, meaning that they needed to recruit fewer neurons to do an easy manual task relative to controls. In essence, musical training also recruits motor networks to allow musicians to engage in the complex motor movements necessary to play music. Recent neuroimaging studies show that visual areas of the brain are activated when people are listening to music, consistent with a number of studies linking auditory and visual cortices (Liang, Mouraux, Hu, & Iannetti, 2013). It may be that listening to music invokes thoughts that invoke visual images, which we know are produced in the visual areas of the brain. Certainly, this is often the goal of some composers, that is, to cause us to bring to mind a particular image. If this is the case, then we may find more interactions between vision and music that might be intuitive. To test this view, Landry, Shiller, and Champoux (2013) compared listeners under normal conditions with those who had been deprived of visual stimulation for 90 minutes. The listeners who had been kept in the dark showed a temporary improvement in their perception of harmo- nicity, that is, whether a chord was in tune or slightly out of tune. The participants who had been visually deprived performed better at this task for up to 5 minutes after the visual deprivation ended. Landry et al. interpreted this to mean that the visual system may play a role in music perception, as there was an observed interaction. Synesthesia The interaction between music and the visual system is even more pronounced in people with color–music synesthesia. Synesthesia is defined as a condition in which a stimulus in FIGURE 13.16 fMRI While Listening to Music The subject (lying in the MRI scanner) listens to instrumental music. Image shows brain areas (mostly the auditory cortex) responding to a meaningful auditory stimulus (beyond the sounds of the MRI apparatus). Synesthesia: a condition in which a stimulus in one modality consistently triggers a response in another modality 386 Sensation and Perception one modality consistently triggers a response in another modality. Estimates of the incidence of synesthesia suggest that it occurs in approximately 1% to 4% of the population (Simner et al., 2006). Synesthesia includes a number of different kinds of cross-modality experiences. For some people with synesthesia, particular words or letters may elicit particular colors, whereas for others, visual stimuli may trigger a taste experience. Color–music synesthesia occurs when particular pitches, notes, or chords elicit experiences of particular visual colors (Figure 13.17) (Farina, Mitchell, & Roche, 2017). Whereas most of us do not have synesthesia, we can do cross-modality matching with some degree of consistency and accuracy. That is, most people will use similar principles to match pitch to color or loudness to temperature (you can try this for vision–audition comparisons in ISLE 13.16). Although normal people may make similar judgments, we seldom experience color while listening to music. The interesting aspect of synesthesia is that these people do experience a sensation in another modal- ity. Note that people with synesthesia are not hallucinating— they are well aware that the secondary experience is illusory. Nonetheless, the experience in the second modality may be vivid and strong. Color–music synesthesia has been described by many musicians and composers, including classical compos- ers Leonard Bernstein and Nikolai Rimsky-Korsakov and jazz pianist Marian McPartland. Recent neuroimaging studies confirm that people with synesthesia have different brain organization than those who do not have synesthe- sia (Farina et al., 2017; Hubbard, Brang, & Ramachandran, 2011; Loui, Zamm, & Schlaug, 2012). These studies show that people with synesthesia tend to have stronger connections between one sensory area and another sensory area than do people without synesthesia. To be more specific, this means that the white matter (axons) between one perceptual area and another is stronger in those with synesthesia. Zamm, Schlaug, and Eagleman (2013) examined the brains of people with color–music synesthesia. They found that relative to controls, people with color–music synesthesia had stronger con- nections between the visual and auditory cortices and areas in the frontal lobe. They found that a tract from the sensory areas to the frontal lobe called the inferior fron- to-occipital fasciculus was enlarged in people with synesthesia (Figure 13.18). Zamm et al. showed that connections between visual areas in the occipital lobe and auditory association regions in the temporal lobe may be differently structured in people with color–music synesthesia. The Neuropsychology of Music Amusia is a condition in which brain damage interferes with the perception of music, but does not otherwise interfere with other aspects of auditory processing. Amusia usually is acquired after brain damage, but there is also a form called congenital amusia, in which individuals are seemingly born with an impairment in music perception. The critical deficit in most forms of amusia, including congenital amusia, is that people with this condition have an impaired ability to discriminate pitches, which affects their music perception but, in most cases, leaves speech perception intact. This is likely the case because pitch is less critical in phoneme perception, but the condition interferes with their ability to perceive and therefore appreciate music (Peretz & Hyde, 2003). Amusia garnered public attention because of the publication of Oliver Sacks’s bestselling 2007 book Musicophilia, which describes a number of fascinating cases of amusia and also describes its opposite, that is, A nd y Zi to /I llu st ra tio n W or ks /G et ty Im ag es FIGURE 13.17 Color–Music Synesthesia Color–music synesthesia: a form of synesthesia that occurs when particular pitches, notes, or chords elicit experiences of particular visual colors Congenital amusia: a condition in which people are inherently poor at music perception ISLE 13.16 Cross-Modal Matchings as a Simulation of Synesthesia 387 Chapter 13: Music Perception cases of people who have become intensely musical after brain damage. Isabelle Peretz and her colleagues at the University of Montreal have been conducting extensive studies on congenital amusia (e.g., Moreau, Jolicoeur, & Peretz, 2013; Wilbiks, Vuvan, Girard, Peretz, & Russo, 2016). People with congenital amusia show deficits in music per- ception as well as production (e.g., they cannot sing in tune and have difficulty learning to play musical instruments). Although it is extremely rare, Peretz estimates that as many as 4% of the popu- lation may suffer from some form of amusia. You may know some of these people as the people who cannot sing even a simple tune and who have no interest in going to concerts with you. Congenital amusia seems to be related to impaired pitch dis- crimination, so people with congenital amusia will not show deficits in speech perception or in most other aspects of auditory perception. As indicated earlier, it is likely that poor pitch perception lies at the heart of congen- ital amusia. However, Peretz (2013) was concerned that an initial deficit in pitch per- ception may spiral in people with amusia. Because they cannot appreciate melodies like normal individuals, they may avoid listening to music. Because of this lack of exposure, their musical impairment will grow with lack of exposure. To remedy this potential in the real world, Peretz conducted experiments in which people diagnosed with amusia agreed to listen to music for an hour per day over a series of weeks. Peretz showed that patients with congenital amu- sia who are exposed to music over a long period of time do not show any improve- ment in pitch perception or in musical understanding (Mignault-Goulet, Moreau, Robitaille, & Peretz, 2012). Thus, mere exposure to music does not “treat” amusia. However, Wilbiks et al. (2016) showed some improvement in music production with an amusic after intensive training. Because of this, Peretz (2013) is con- vinced that congenital amusia is a genet- ically linked syndrome that occurs in some people. Indeed, Peretz (2008) described research that shows that congenital amusia occurs within families more than chance would predict, thus pointing to a genetic component. She argues for a neuroanatomical pathway that might be suppressed or impaired in people with congenital amusia. For example, she argued that research shows that there may be deficits in the transmission of information from the auditory associative cortex, that is, those areas that surround FIGURE 13.18 The Brain in Synesthesia (Zamm et al., 2013) People with color–music synesthesia have stronger connections between the visual and auditory cortices and areas in the frontal lobe. This is demonstrated in these functional magnetic resonance images. Gene 1 Tonal encoding of pitch Failure to detect anomalous pitches in melodies Indifference to dissonance Failure to recognize tunes Singing out of tune Acoustical encoding of pitch Gene 2 Gene 3 Env. 1 Env. 2 Env. 3ETIOLOGY BRAIN COGNITION BEHAVIOR FIGURE 13.19 Causation of Congenital Amusia This diagram illustrates the potential causes and manifestations of amusia. 388 Sensation and Perception the primary auditory cortex, to areas in the frontal lobe, such as the inferior frontal gyrus (Figure 13.19). We will cover one last interesting aspect of congenital amusia. We have seen that one of the primary deficits is an inability to distinguish close pitches. Because of this pitch confusion, it may be hard for individuals to distinguish music that is in tune from music that is out of tune, and music that is consonant from music that is dissonant. Thus, most individuals with congenital amusia tend to shy away from music. However, what hap- pens to individuals with congenital amusia when they attempt to learn a language such as Mandarin (Chinese) or Vietnamese, in which the pitch with which a word is said is important to meaning? Early studies suggest these are extremely difficult languages for congenital amusia individuals to learn as second languages, but we do not know if native Mandarin speakers with congenital amusia have a deficit in understanding semantics conveyed by pitch (see Peretz, 2008). Future research will hopefully tease these issues out. TEST YOUR KNOWLEDGE 1. What neural differences does one see in professional musicians compared with appropriate control populations? 2. What is congenital amusia? Why might its incidence in the population be underestimated? LEARNING, CULTURE, AND MUSIC PERCEPTION 13.3 Discuss how learning and culture affect music perception. Music and Language One of the ongoing debates in the field of music perception is the extent of the meta- phor between language and music. Is music a form of language in which the ideas trans- mitted are not words and semantic meaning but notes and emotions? Are there special parallels in the processes that allow us to understand language and appreciate music? Some researchers argue that language and music are very similar processes, whereas other researchers argue for little overlap between the two. We briefly address some of the arguments for both views here. One of the most passionate spokespeople for the idea that language and music share similar neurocognitive systems is noted music neuroscientist Aniruddh Patel (Patel, 2008, 2013). Patel asserts that language and music have much in common at the behav- ioral level. First, music and language are both perceptive (listening) and productive (singing, talking) systems in which perceiving and producing are equally important (if only to sing in the shower). Moreover, both involve the perception of novel and com- plex sounds that unfold rapidly over time. Subjectively, hearing a melody is different from hearing a sentence, but both are sound stimuli that transmit meaning, so it is not unrealistic to expect some overlap. Patel (2013) also argues that both music and lan- guage have structure that must be followed for the sounds to make sense to listeners. For example, language has syntax that governs which words can be joined together to make a coherent sentence. Patel argues that music theory describes a syntax that serves 389 Chapter 13: Music Perception a similar function in music, namely, limiting those notes that can be joined together to form consonant music. Similarly, words have specific meanings, but that meaning can depend on context. For example, the word bugger may be a term of affection in one situation but a vile insult in another. Meaning can also vary in music as a function of context. In many situations, minor keys denote sadness, but there are also a great many wedding celebration songs written in minor keys. On a neural level, there are also some striking parallels between music and lan- guage. First, both use the same basic auditory machinery. Whereas language uses more neural space in the left hemisphere, music perception and production appear to be housed in the analogous regions of the right hemisphere. Moreover, the better someone becomes at language, the more right hemisphere involvement we see. Similarly, many studies have shown the acquisition of musical expertise is accompanied by greater left-hemisphere involvement in music (Patel, 2013). Nonetheless, there are also important differences between music perception and lan- guage perception that must be acknowledged, which cloud the analogy between music and language. First, speech perception is based on the inference of subtle differences in the patterns that produce different vowels and consonants, whereas music focuses on pitch and pitch contrasts. We can say a sentence without varying the pitch; the move- ments of the mouth create different sounds that carry phoneme information. Similarly, music can be sung using different phonemes, but as long as the pitch remains, we rec- ognize the music as such. For example, think of the conventional way of singing Henry Mancini’s “Pink Panther” (1963)—“dead ant, dead ant, dead ant, dead ant, dead ant, dead ant, dead ant. . . .” We could certainly change the semantic content from dearly departed insects to anything we like and still represent the melody of this song. Indeed, when humming the melody, we may think of the Pink Panther’s bumbling French detec- tives or cool jazz, but seldom do we think of it as a dirge for the Formicidae. Thus, the meaning in music extends beyond what is sung. Finally, it is also possible to interpret the neural evidence as suggesting differences between language and music. That is, lan- guage predominates in the left hemisphere, whereas music predominates in the right hemisphere. What could be a more basic difference than that? Culture and Music Perception An obvious truism about music is that it varies so much—from culture to culture, from generation to generation, from “pop” traditions to “highbrow” traditions. Given this incredible diversity of music, are we justified in making the generalizations about music perception that we have been making throughout this chapter? Our assertion here is that despite the differences in music across cultures, there are some universalities that allow us to talk about music and not just “varieties of music.” We have already dis- cussed, for example, the universality of the octave, which all music traditions respect. All cultures use pitch and rhythm to express emotion in their music, either with or without singing. Therefore, we now briefly consider some rules that may be universal and some that may be specific to our own Western traditions. A brief reminder of the context of the term Western music: Our use of the term refers to a huge gamut of music, including what most of us would commonly call “classical” music and “pop” music, as well as jazz, hip-hop, rap, reggae, southern rock, rock ’n’ roll, country, grunge, and so on. All of these styles follow the Western music tradition. However passionately you may love one form (e.g., reggae) and hate another (e.g., country), they all share the Western music tradition and therefore have more in common with one another than they do with non-Western forms of music. Western musical styles use the same scale structure, the same relations among notes within the octave, a common means of notating written music, and a common set of assumptions 390 Sensation and Perception about what is consonant and what is dissonant. Indeed, they use many of the same musical instru- ments (Figure 13.20). Although from where most of us sit, the tattooed and nose-ringed fans of a heavy metal group may have little to do with the tuxedoed and gowned attendees of an opera, they are both engaging with music that derives clearly and directly from Western music traditions. But outside the Western music tradition, we find music that is organized by radically different principles. For example, the Indian rag (or raga) scales that gov- ern much traditional Indian music are very different from Western scales. First, in much Indian music, there are 22 notes within each octave, far more than the traditional Western 12 (ISLE 13.17). In addition, few of these notes fall exactly at the same frequen- cies as Western notes (Figure 13.21). Thus, at a basic level, Indian music is using notes that would fall between the notes of the Western music scale. It would be impossible to play raga music on a Western instrument, such as a piano or any woodwind instrument. Moreover, in traditional Indian music, different scales are associated with different times of year, different moods, and even different times of day (Tan et al., 2010). Research shows that the networks in the brain react differently to daily exposure to Indian music as opposed to Western music (Ambady & Bharucha, 2009). Similarly, Javanese gamelan music uses different scale systems, called slendro, which is a five-tone scale, and pelog, which is a seven-tone scale. These scales bear a rough cor- respondence to Western scales, but the notes are distributed at different intervals relative to the octave of each scale. Perlman and Krumhansl (1996) found that Western listeners were impaired in their ability to detect aspects of Javanese music relative to those who were more familiar with it. Performance may also vary across musical traditions. In the Western tradition—from art and classical music on one hand to garage rock bands on the other—we have perform- ers, and we have listeners. Listeners may dance and shout in less formal venues (but do not try that in a symphony concert hall), but they are not part of the musical perfor- mance. This history of performers and listeners goes back centuries in our musical traditions. In many non-Western music traditions, it is expected that all present will be part of the act of making music as well as listening to it. Musical traditions differ in their approaches to rhythm as well. In particular, research has focused on differences between Western musical approaches to rhythm and those of traditional West African drumming (Temperley, 2000). Temperley argues that syncopa- tion (varying the emphasized beat in a musical piece) is usually more pronounced in West African traditions (Figure 13.22). However, syncopation became a part of the Western tradition in jazz and in classical music in the 20th century, perhaps borrowing from African traditions. Thus, differences in rhythm may be quantitative changes among simi- lar traditions rather than markers of completely different musical traditions, as we see in the case of different notes within a scale. © iS to ck ph ot o. co m /N ad ez da st oy an ov a FIGURE 13.20 Havana, Cuba, May 10, 2013 The conductor performs with a brass band in the streets of Havana, Cuba, in Central Park Square near Hotel Inglaterra. ISLE 13.17 Indian Raga Music Example Bilawal scale Charukesi scale Poorvi scale FIGURE 13.21 Indian Raga Scales The Bilawal scale is identical to our major scale, but the other ragas are quite different from the scales used in Western music. 391 Chapter 13: Music Perception An interesting pattern that may be universal across musical traditions is the relation between music and the experience of emotion. Seemingly, in all cultures, music is used to convey emotion ranging from sadness and anger to joy and ecstasy. Most of us know this experience firsthand. A song comes on the radio and reminds us of a partic- ularly romantic night with our significant other; another song reminds us of a particularly unpleas- ant breakup. The composer Johann Sebastian Bach wrote the sarabande of his Violin Partita No. 2 in D minor (Opus 1004) in the weeks after the death of his wife. Even today, more than 300 years later, we can hear the inconsolable sad- ness in the piece (you can hear a sample of this piece on ISLE 13.18). Even if you have never listened to classical music, you should be able to determine the sadness conveyed by the minor chords and the haunting melody. However, moving across cultures, it may be difficult to detect the intended emotion in a musical composition. In the Western music tradition, minor scales often convey sadness, as certainly Bach’s Violin Partita No. 2 in D minor does. However, in Middle Eastern music, those same minor scales may be used to express joy. The research, however, suggests that listeners can detect the intended emotion across musical traditions, supporting the idea of the universality of emotion in music. For example, Balkwill and Thompson (1999) found that Western listeners could accurately identify the intended emotion in Indian raga music. Listeners attended to pitch (lower is sadder, higher is happier) and tempo (slower is sadder, faster is happier) to find the intended emotion, and these features appear to be universal. Meyer, Palmer, and Mazo (1998) showed that Western listeners could identify the emotional content of traditional Russian laments by listening for timbre cues in the singers’ voices. Thus, it may be that the expression of emotion in music transcends musical traditions. TEST YOUR KNOWLEDGE 1. What are the features in common between language and music? Do you think they evolved for similar purposes or different ones? 2. What are minor scales? What do they represent in different cultures? ISLE 13.18 Bach’s Violin Partita No. 2 in D minor © iStockphoto.com /arrow sg FIGURE 13.22 Santa Maria, Cape Verde, Sal, March 27, 2016 EXPLORATION: Musical Illusions Our perceptual systems are designed not only to detect stimuli in the environment but also to extract meaning from those sensory stimuli. In a sense, this is what music is—the extraction of patterns and emotional meaning out of a stream of patterned auditory stimuli. As we have seen throughout this chapter, what creates music is a pattern in melody, in harmony, or in rhythm. And as you have seen throughout this book, when our perceptual systems attend to patterns, they can be tricked into perceiving patterns even when elements of those patterns are missing. Music is no exception. Music perception researchers have identified and created a number of engaging musical illusions. These illusions are fun to listen to, but they also tell us about the underlying structures and functions of musical perception. 392 Sensation and Perception We will start with one of the more compelling and frus- trating musical illusions, called Shepard tones after the researcher who first designed this illusion (Shepard, 1964). Shepard Tones In Shepard tones, one hears a scale that sounds as if it increases in pitch continually. (The illusion can also be designed so that one hears a scale that sounds as if it is decreasing in pitch.) Each sound in the scale seems a bit higher than the one preceding it, but the listener eventually realizes that the tone one is hearing is back at the pitch at which the scale started, although all one hears is increasing pitches (the example on ISLE 13.19 is a must-hear). That is, the pitches sound as if they are getting continually higher, but in fact, the sound fre- quencies return back to lower frequencies without our noticing it in any change from one note to the next. The illusion can also be run in reverse, with the per- ception being that the notes get lower and lower, when actually they do not. Again, please listen to t h e d e m o n s t r a t i o n — in this case, hearing is believing. How did Shepard do this? The illusion is created by simultaneously sweeping dif- ferent pure tones that are an octave apart. By sweeping, we mean starting at one note and sliding through all the intermediate frequencies to another note. Thus, the scale may start with an A at 220 Hz, and the tone slowly slides up to an A at 440 Hz. At the same time, there is another A starting at 880 Hz and sliding down to 440 Hz. This creates the illusion of a rising pitch, even though there is no dif- ference in the frequencies being presented (Figure 13.23). In essence, this is an auditory ver- sion of the barber pole illusion, in which the color pattern looks as if it is continually going up or going down, despite the obvious fact that this cannot be occurring (Figure 13.24). The Octave Illusion Deutsch (1974) described the octave illusion, a phenomenon she had been studying in her laboratory. You can listen to the octave illusion in ISLE 13.20. It is a stereo illusion, so please make sure that you listen to it with headphones on. If you do not, the illusion will not work properly. In the octave illusion, one tone is presented to one ear while another tone, exactly one octave higher or lower, is pre- sented simultaneously to the other ear. However, the next note combination is of the same two notes, but to the opposite ears. That is, if a middle G is presented to the left ear, and the G an octave lower is presented to the right ear, the next notes will be a middle G presented to the right ear and the G an octave lower presented to the left ear (see Figure 13.25). This is an illusion, so think first about what you should hear: the same note, alternating between ears, and another note also alternating between ears. It should sound the same, as the stimuli are the same for both notes, just presented to different ears. But listen to it on ISLE again. What do you actually hear? What most people report hearing is the fol- lowing: You hear a single note (the middle G) in your right ear followed by a single note an octave lower in your left ear. And then a continuous alternation between the two occurs, regardless of which ear the higher G is actually being pre- sented to. So what you hear differs from what is actually being presented—hence the term illusion. If you reverse your ISLE 13.19 Shepard Tones FIGURE 13.23 Spectrogram of Shepard Tones The spectrogram shows that the interlacing of rising tones along with an octave shift down (indicated by yellow arrows) causes the perception of a gradually increasing sequence in pitch, even though the actual sequence does not rise in pitch. ISLE 13.20 Octave Illusion FIGURE 13.24 The Barber Pole Illusion When this barber pole is set in motion, it looks as if it continually rises (or falls), when it actually just circles back to the same place. 393 Chapter 13: Music Perception headphones, you get the feeling not that the sounds are com- ing from different ears, but that the order of notes reverses (Deutsch, 1975a). Some left-handers may hear the pattern in reverse, that is, the higher note in the left ear. This illu- sion is likely due to differences in pitch processing between the left and right auditory cortices (see Deutsch, 2013). You can also go to Dr. Deutsch’s website (http://deutsch.ucsd.edu/ psychology/pages.php?i=201) to hear a number of variants of this illusion. The Scale Illusion Deutsch also discovered the scale illusion that shares several features with the octave illusion (Deutsch, 1975b). Like the octave illusion, the scale illusion has a different pattern presented to each ear (Figure 13.26). You should go to ISLE 13.21 to try what you would hear in each ear separately following the directions on the ISLE. Each pattern sounds rather random and violates many of the gestalt principles of hearing mel- odies that were discussed earlier. Next, try listening with headphones on both ears and play the sequence. Most peo- ple hear the patterns shown at the bottom of Figure 13.26. A descending and ascending scale is heard in one ear and an ascending and descending scale is heard in the other ear. In other words, each ear hears a scale with the direction of the scale in the opposite direction for each ear. You can figure out how the scales can be assembled from what is played in each ear if you start with the first note for the right ear. With this note you are at the top of a scale. The next note for this descending scale is found in the second note played but in the left ear. The third note is back in the right ear, and so forth. Refer back to the role of the gestalt principles in the perception of melody discussed earlier. For example, proximity and good continuation are important prin- ciples in hearing a melody amid a whole constellation of notes. The tunes played in each ear violate both of these gestalt principles, but our perception follows both of them very nicely. Compare the two parts of Figure 13.26 and see how the perception follows the gestalt laws much better than what is played in each ear. It seems our need for coherent perception overrides what ear the sound arrives in, even when the sound is limited to one ear as when head- phones are used. ISLE 13.21 Scale Illusion Sound pattern Sound perception = right ear= left ear FIGURE 13.25 The Octave Illusion Deutsch (1974) presented one tone to one ear and another tone, exactly one octave higher or lower, simultaneously to the other ear. However, the next note combination was of the same two notes, but to the opposite ears. You hear a single note (middle G) in your right ear, followed by a single note an octave lower in your left ear. Fr om h tt p: // de ut sc h. uc sd .e du /p sy ch ol og y/ pa ge s. ph p? i= 20 3 FIGURE 13.26 Scale Illusion = left = right = 240 (a) Sound Pattern (b) (c) Perception From http://deutsch.ucsd.edu/psychology/pages.php?i=203 394 Sensation and Perception C C# A# G# F# F E G A D# B D FIGURE 13.27 The Pitch Class Circle (Deutsch, 1986) As you move to the left on the circle, notes sound higher in pitch. As you move to the right, notes sound lower in pitch. ISLE 13.22 Tritone Paradox The Tritone Paradox This illusion was also discovered by Deutsch (1986). In music, tritone refers to the half octave, or the interval spanning six semitones. Thus, in the key of C major, there is one tritone: If you start on F, you can go up six semitones to B. Similarly, E and A-sharp are tri- tones (Figure 13.27). In the tritone paradox, Deutsch presents stimuli generated similar to the way Shepard generated his paradox- ical scale; that is, each note is an envelope of sound sweeping from one octave to the next, but with a heard pitch equiva- lent to the lower note. Thus, in the tritone paradox, Deutsch played a note with a perceived pitch of C and one with a perceived pitch of F-sharp, a tritone away (you can hear this illusion in ISLE 13.22). Here’s the paradox: Some peo- ple hear the notes as ascending, as in a lower C to a higher F-sharp, whereas other people hear the notes descending, as in a higher C to a lower F-sharp. Deutsch even tested musi- cians and found the same result. Thus, it is good to do this demonstration in a group, because people will disagree on what they just heard. What is the explanation of the individual differences in the perception of these tritones? Deutsch and her colleagues have studied it extensively (Deutsch, 2013). It turns out that there are regional differences in the perception of the tritone par- adox. One sees a different distribution of hearing the tritone paradox as ascending or descending whether one is examining Americans in California or the English in England. Vietnamese who emigrated to the United States at an early age show a different distribution of perceptions of the tritone paradox than do Vietnamese who emigrated later. Deutsch thinks these differences have to do with the pitches typically heard in speech for different communities. Vietnamese and the English show greater variance in their pitch patterns in speech than do Americans. Thus, this illusion potentially shows a relation between music perception and speech per- ception. For more on Deutsch’s work, you can visit her website or listen to a number of other of her illusions at http://www .radiolab.org/story/292109-musical-illusions (also linked to on ISLE 13.22). APPLICATION: Music Perception in Hearing-Impaired Listeners Most everyone knows the story of the great composer, Ludwig van Beethoven (1770–1827) (Figure 13.28). Beethoven was a composer, conductor, and piano virtu- oso. He defined the transition from the Classical era to the Romantic era in music. However, Beethoven started suffer- ing from hearing loss in his late 20s. By the age of 40, he was sufficiently hearing-impaired that he was forced to stop per- forming and conducting. His knowledge of music, his work ethic, and his ability to create musical imagery allowed him to continue composing music even after he could no longer hear what he was writing. The story of Beethoven is usually considered uniquely tragic, given the man’s dedication to music and his subsequent deafness. However, his story is not unique. Many music lovers experience hearing loss, and many musicians experience hearing loss, sometimes as a consequence of their careers as musicians. Indeed, profes- sional musicians have almost 4 times the risk of hearing loss compared with people in the general population (Schink et al. 2014). Many popular musicians today experience hearing loss, including Peter Townshend, Ozzy Osbourne, Neil Young, Sting, Phil Collins, Grimes (Claire Boucher), and will.i.am (of the Black Eyed Peas). These musicians are 395 Chapter 13: Music Perception all “rockers,” but hearing loss also occurs among classi- cal musicians (Toppila, Koskinen, & Pykkö, 2011). Other famous composers such as Bedřich Smetana, Gabriel Fauré, and Ralph Vaughan Williams all suffered from hearing loss. In many musicians, the hearing loss may result from pro- longed exposure to loud music over a long time. The high decibel level of the sound can cause hearing loss, particu- larly in the high-frequency range. This may seem obvious in the case of the rockers, as this form of music is often played very loudly and with electronic amplification. Classical and jazz musicians may also be exposed to loud music, both from their proximity to their own instrument and their proxim- ity to instruments close by. For example, violinists tend to show selective hearing loss in their left ears, which is close to the source of sound in their violin (Royster, Royster, & Killion, 1991; Schmidt, 2011). In Beethoven’s time, there was very little that could be done about hearing loss. Although Beethoven had some residual hearing, he used a variety of “horns” as hearing aids. The problem with these horns is that they amplified all sounds equally, much as early hearing aids did in the past. For a musician, this only makes the problem worse because being a musician is as much about hearing differences among sounds as it is hearing them at sufficient amplitude. A collection of Beethoven’s hearing aid horns can be found at the Beethoven museum in Bonn, Germany, if your travels ever take you there (Ealy, 1994). Today, however, musicians can continue to perform with the help of digital hearing aids, and some profoundly deaf individuals can have cochlear implants adjusted to allow a returned sense of musical enjoyment. Sensorineural hearing loss is likely to affect the hearing of some frequencies more than other frequencies. For musicians exposed to loud music, it is often the high frequencies that are most impaired. Digital hearing aids can selectively amplify the impaired range of frequency. Moreover, hearing aids can be selectively programmed to increase the amplitude of musi- cally relevant stimuli (Chasin & Hockley, 2014). But increased amplitude is only the first part of the equation. Music is played louder than speech and at a greater range of frequency than most speech, so ideally a hearing aid wearer will have the flexibility in his or her hearing aid’s programming to have separate programs for conversation and for music listening. The listener will want to hear both the low-frequency sounds of the bass and the high-frequency sounds of the lead singer’s voice, hence the need for amplification at a greater range of frequencies than for speech. In addition to the amplification, the fidelity or sound quality is also crucial to the enjoyment of music but less relevant for speech perception. Timbre per- ception is also affected by hearing loss, as high-frequency components contribute to timbre. If those frequencies cannot be heard, it interferes with the completeness of the music. As such, improving the timbre of sound heard through hearing aids is an issue that needs improvement. One problem with the sound quality achieved while using hearing aids is that the processing of the sound signal by the aids can slow down the transmission of information to the cochlea. For example, one problem any hearing aid wearer struggles with is the feedback problem that often occurs espe- cially when loud sounds are in the environment or when there is physical contact with the aid. When the hearing aid receiver picks up the sound the transmitter is projecting, a characteris- tic whistle occurs, which interferes with all sound perception. Digital hearing aids have cancelation systems to counteract feedback, but these may slow down processing, leading the listener to a smeared or blurred sound rather than a high-clar- ity fidelity (Chasin & Hockley, 2014). Although many listen- ers of music might not notice the difference in sound quality, musicians often do because their training leads them to be more sensitive to this factor (Zakis, Fulton, & Steele, 2012). As complex as hearing aids can be, they only amplify sound and cannot change the frequency tuning of the individual. With hearing aids, it is still the hair cells of the cochlea that are actually doing the hearing. When hearing loss affects the outer hair cells, the frequency tuning of the inner hair cells becomes greater, meaning the same inner hair cell is responding to a FIGURE 13.28 Ludwig van Beethoven Joseph Karl Stieler [Public dom ain], via W ikim edia Com m ons 396 Sensation and Perception range of frequencies (Moore, 2012). This means that no mat- ter how good the amplification or the tuning from the hearing aid is, there is going to be a loss of sound quality because people with sensorineural hearing loss cannot discriminate among close pitches (or chroma, in musical terms), thus lead- ing to a less clear sound (Chasin & Hockley, 2014). An issue for musicians is that most hearing aids automati- cally dampen sound at loud volumes. This is problematic for musicians who must be able to hear the correct dynamics, as played by themselves and musicians in their group. The dampening of the sound in this way may cause the quality of the sound to be affected as well. This leads to an odd prob- lem for musicians with hearing loss because it is their hearing aid’s design to protect them from loud sounds (which may have caused their hearing loss in the first place) that is inter- fering with their ability to play the music correctly. Chasin and Hockley (2014) review a number of solutions that musi- cians may use to help them with this problem, including simply putting tape over the hearing aid so a louder level of ambient sound is necessary to trigger the dampening. Older musicians may also switch the instrument they are playing to one whose range of sound is less in their impaired range. Because high-frequency sounds are often the first ones lost, violin players may find they can continue to play music if they switch to the viola. Clarinet players can focus on their skill as bass-clarinet players. Chasin and Hockley also review recent technological advances that will allow musicians to safely hear these loud sounds without distortion. Next consider individuals with cochlear implants who wish to enjoy music again (or for the first time) (Figure 13.29). Many of the problems that confront hearing aid users are compounded in cochlear implant users. As with hear- ing aids, the main function of cochlear implants is speech perception (Limb & Roy, 2014). If “location, location, location” is the mantra of real estate agents, then “speech perception, speech perception, speech perception” is the mantra of audiologists (we’ve been waiting all book to use that line). Speech perception involves a much more limited band of frequencies and a much narrower range of loud- ness. So programs that work well for speech perception do not necessarily work well for music listening. The biggest single difficulty for cochlear implant users is that pitch per- ception is very poor. Because of the tuning characteristics of the implanted electrodes, there is just insufficient ability to distinguish between notes. Because the notes are hard to distinguish, cochlear implant users have a very difficult time determining basic musical features such as melody and harmony (Limb & Roy, 2014). This leads to even further difficulties in perceiving musical features such as the per- ceptual integration needed for auditory stream segregation. However, Limb and Roy claim that with improving technol- ogies and much training, it is possible for cochlear implant users to regain some ability to perceive and therefore enjoy music. In sum, even people with profound hearing impair- ment can still find enjoyment in the art of music. FIGURE 13.29 Music and Cochlear Implants Ja m es K in g- H ol m es /S ci en ce S ou rc e CHAPTER SUMMARY 13.1 Explain how frequency is related to pitch, chroma, and the octave. Music is ordered sound made and perceived by human beings created in meaningful patterns. Humans have been making music since at least the Stone Age. Pitch is the subjective experience of sound that is most closely associated with the frequency of a sound stimulus. Pitch is related to the experi- ence of whether a sound is high or low, such as the notes at the far right and far left of a piano. The octave is the interval between one note and a note with either double the frequency or half the frequency of that note. Chroma is the subjective quality of a pitch. We judge sounds an octave apart to be of the same chroma. In Western music, there are 12 semitones or 12 equivalent intervals or notes within each octave. Western music also uses an equal-temperament scale in which the difference between each successive semitone is constant in pitch and a constant ratio in frequency. Harmony occurs when two or more notes sound pleasant when played together. In musical terms, consonance is the perception of pleasantness Chapter 13: Music Perception 397 or harmony when two or more notes are played; that is, the notes fit with each other. In contrast, dissonance is the per- ception of unpleasantness or disharmony when two or more notes do not fit together. Pitch is not the only critical aspect of music. Dynamics refers to the relative loudness and how loudness changes across a composition. Rhythm is the temporal patterning of the music, including the tempo, the meter, and the beat. Tempo is the speed at which a piece of music is played. Meter is the tempo- ral pattern of sound across time. Beat refers to spaced pulses that indicate if a piece is fast or slow. Timbre is the complex sound created by harmonics. Attack is the beginning buildup of a note. Decay refers to how long the fundamental frequency and harmonics remain at their peak loudness until they start to disappear. All of these factors are critical to music. Melody is the rhythmically organized sequence of notes, which we perceive as a single musical unit or idea. A scale is a set of ordered notes starting at one note and ending at the same note one octave higher. Key refers to the tonic note (e.g., C in a C major or minor scale) that gives a subjective sense of arrival and rest in a musical piece. Transposition allows more than one version of the same melody, beginning on dif- ferent notes but containing the same intervals or sequences of changes in notes. Gestalt principles predict many of the patterns that determine melody in musical compositions. 13.2 Summarize the basic neuroscience of music, including how training and experience can affect the representation of music in the brain. The neuroanatomy of music is a bit different than the neuro- anatomy of many other cognitive abilities because music is in some ways an optional feature of human cognition. Most prominent are areas in the right temporal cortex, adjacent to the right auditory regions. One clear finding is that music per- ception usually causes greater activation in the right tempo- ral lobe than the left temporal lobe, but we also see activity in the left temporal lobes and the frontal lobes when perceiving music. In trained musicians, it has been found that the level of activation of the left hemisphere increases when listen- ing to music. Synesthesia is a condition in which a stimulus in one modality consistently triggers a response in another modality. Color–music synesthesia occurs when particular pitches, notes, or chords elicit experiences of particular visual colors. People with color–music synesthesia show a greater activation of the tract that connects the frontal lobes to the auditory cortex. Amusia is a condition in which brain damage interferes with the perception of music but does not otherwise interfere with other aspects of auditory process- ing. Congenital amusia is a condition in which people are inherently poor at music perception. There is some research that suggests connections between music and language. 13.3 Discuss how learning and culture affect music perception. Music is seen in all cultures, but the music of other cul- tures differs in systematic ways from Western music. For example, ragas from India use a 22-note scale instead of the Western 12-note scale. Moreover, the notes of the raga scale fall at different frequencies than Western notes. Shepard tones, the octave illusion, and the tri- tone paradox are all illusions that illustrate how listeners extract musical meaning from stimuli, even when it is not an accurate description of the physical stimuli. REVIEW QUESTIONS 1. Provide a definition of music. How is musical per- ception different from other forms of perception, and how is it similar? 2. What is the relation of pitch to chroma? What is the octave? And what evidence is there that the octave is a musical universal? 3. What is meant by consonance and dissonance? How do they relate to chroma and octaves? 4. What are the differences between tempo, meter, and beat? How does each contribute to musical perception? 5. What is timbre? What defines the physical differ- ences that make up timbre? How is timbre used by musicians to convey meaning or mood? 6. What is a scale? How does it differ from culture to culture? What starts and ends a scale? What is meant by transposition? 7. What areas of the brain are critical to perceiving music? What other areas are needed for musicians? How does the brain change with musical training? 8. What is synesthesia? What is color–music synesthesia? What is amusia? 9. How does music differ from culture to culture? In particular, how does the scale system differ between Western music and ragas from India? 10. Describe two musical illusions. What are the physi- cal stimuli? How are those stimuli perceived? Sensation and Perception398 PONDER FURTHER 1. Music perception necessarily involves experience. How can researchers distinguish between auditory processes that are learned versus those that may be “built” into the brain by studying music? 2. Much of what we have been discussing in this chapter is the perception of music rather than the production of music. Performing musicians are engaged in physical movements. How might such movement affect the per- ception of music and why? KEY TERMS Attack, 379 Beat, 378 Chroma, 375 Color–music synesthesia, 386 Congenital amusia, 386 Consonance, 377 Decay, 379 Dissonance, 377 Dynamics, 377 Equal-temperament scale, 376 Harmony, 377 Key, 381 Melody, 380 Meter, 378 Music, 373 Octave, 375 Rhythm, 377 Scale, 380 Semitones, 375 Synesthesia, 385 Tempo, 377 Transposition, 381 Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review, study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning. Learning Objectives Digital Resources 13.1 Explain how frequency is related to pitch, chroma, and the octave. 13.2 Summarize the basic neuroscience of music, including how training and experience can affect the representation of music in the brain. Musical Disorders: From Behavior to Genes This Is Your Brain. This Is Your Brain on Music Born to Be Tone Deaf? Ani Patel Talks About Musical Training and the Brain Charles Limb: Your Brain On Improv Daniel Levitin: Music and the Brain What Color Is Tuesday? Exploring Synesthesia 13.3 Discuss how learning and culture affect music perception. What Can Experiments Reveal About the Origins of Music? The Development Of The Aesthetic Experience Of Music: Preference, Emotions, And Beauty Think Before You Clap: You Could Be Beat Deaf florin1961 (Florin Cnejevici)/Alamy Stock Photo 14Touch and Pain Kai Tirkkonen/The Image Bank/Getty Images LEARNING OBJECTIVES 14.1 Classify how touch perception is actually a number of senses served by a number of receptor systems. 14.2 Sketch the two main pathways from receptors in the skin to the somatosensory cortex in the brain. 14.3 Examine the role of endogenous opioids in controlling pain perception. 14.4 Contrast the difference between pain and itch and how they interact. 14.5 Assess the role movement plays in enhancing the perception of touch, allowing us to perceive objects. 14.6 Illustrate how the vestibular system operates to perceive head rotation and movement and interact with other senses to help us maintain balance. INTRODUCTION Think of many of the things that bring us the most pleasure and the most discomfort. Think of the relaxation brought by stroking the fur of a sleepy cat. Think of the feeling of the joy of life when the cool water of a pool rushes around your face as you dive in. Think of the thrill and the excitement as your lips meet those of your romantic partner. You will note that all of these desirable experiences occur through our senses that arise from the skin. Similarly, think of unpleasant sensations, the throbbing pain of a tooth- ache, the quick pain when you cut yourself while shaving, or the maddening itchiness of a mosquito bite. You will note here that many of the experiences we deem to be most negative also occur through the skin senses, mainly the skin sense we know as pain. We often take our senses related to touch for granted, but in reality, they are essential in many ways. Take a look at Figure 14.1 and see if you can imagine how these various surfaces would feel to your skin. Luckily, we have very poor imagery for pain, so it is hard to conjure up the experience of a toothache without actually having one. The touch senses are different in an important way from the already discussed vision and audition. Vision, audition, and olfaction can work at a distance and are sometimes called distal senses. For example, I can see the boat on the lake even though it is a half mile away, I can hear the barking of the dogs even though they are across the street, and ISLE EXERCISES 14.1 Action of Mechanoreceptors 14.2 Mechanoreceptors and Aristotle’s Illusion 14.3 Heat Grille 14.4 Somatosensory Pathways 14.5 Melzack and Wall’s Gate Control Theory 14.6 Professor Ramachandran and Phantom Limb Syndrome 14.7 Treatment of Phantom Limb Pain 402 Sensation and Perception FIGURE 14.1 Somatosensory Experiences Imagine what each of these would feel like on your skin. Some of these may be touch experiences we long for, whereas others may be perceptions we would prefer to avoid. I can smell the cookies baking in the other room. Touch is more intimate; it requires direct bodily contact and preferably contact and bodily motion. Sometimes touch and its related sensations are called proximal senses. As such, touch is interactive. It is often quite diffi- cult to detect whether a surface is smooth or rough using a nonmoving finger. We move our fingers and the rest of our skin to sense surfaces. Thus, from the very beginning of our discussion of the touch senses, we will need to consider movement of the body as well. Here is another thought about the importance of our touch systems. Think about waking up in the middle of the night and feeling very hungry. You do not want to wake anyone else up, so you decide not to turn the lights on. Without visual or auditory cues, you can feel your way from your bedroom into the kitchen, find the cutlery drawer and remove a fork, feel the piece of cake that was left on the counter from dessert earlier in the evening, and enjoy your midnight snack alone in the dark. You do all these tasks that you might normally do with visual guidance by using your touch senses alone. You may move about the house a bit more slowly as you feel your way along the wall, but you can probably do this without stumbling. If you have midnight snacks too often, you may need to go on a diet, but the point is, we can feel our way around familiar environments without the use of sight. Before we begin our discussion of the details of our touch senses, we will note we are using the plural to describe these senses. Properly speaking, there is not a single sense of touch, but several somatosensory senses, that is, those sensations that arise from the skin, muscles, and other interior senses. Just as an aside, soma in the word somatosen- sory means “body,” so these senses are intimately connected to our body and inform our brain about events in and on our body. There are different kinds of receptors in our skin to detect light pressure, deep pressure, pain, coldness, and heat. There are also receptors within our muscles that help us regulate the position of our body, which we also consider in this chapter. This chapter also considers the vestibular system, which is not a somatosensory system, per se, but we examine it here to ensure that we cover © iS to ck ph ot o. co m /W 1z za rd © iS to ck ph ot o. co m /h av es ee n © iS to ck ph ot o. co m /e so lla © iS to ck ph ot o. co m /C Pa ul us se n 403 Chapter 14: Touch and Pain all the sensory systems in this book. In addition, stimulation for this sense results from the motion of our body, in particular, our head, so it fits in a broad way. The vestibular system is rather unique in that, unlike the other senses we consider here, we are not usu- ally aware of its functioning. We also include a diversion into the realm of nonhuman animal senses in this chapter. The Exploration section is about electroreception in fish. THE SKIN AND ITS RECEPTORS 14.1 Classify how touch perception is actually a number of senses served by a number of receptor systems. If you are ever on a game show, and you are asked the following question, you will now get it right and win thousands of dollars: What is the heaviest organ of the body? The answer is the skin, which weighs in at more than 4 kg (9 lb). The skin can be considered the sense organ of touch, much as the eyes house the visual system receptors and the ears house the auditory receptors. Touch receptors exist everywhere on the sur- face of the skin, though not necessarily in a uniform fashion. Touch receptors are closely grouped on your fingers and lips and less closely grouped on your upper back, legs, and arms. There are also touch receptors inside your mouth and on your tongue. The skin is a complex organ com- posed of differing layers, despite its familiar surface features. Many touch receptors are located just below the outer layer of skin, called the epidermis. The epidermis is interestingly avascu- lar, as it draws oxygen directly from the air rather than from the blood- stream. The epidermis is thickest on our palms and the soles of our feet, and it is thinnest on our eyelids. In addition to its sensory functions, the epidermis functions to keep out pathogens and keep in fluids. Indeed, it is composed mainly of dead cells that protect the dermis below. Below the epidermis is the dermis, which houses most of our touch receptors. To help you keep track of terms, epi means “above” or “on top of,” so the epidermis is above the dermis. The dermis also holds the connective tissue and has a blood supply. Skin can also be divided into hairy skin, that is, skin with hair growing on it, and glabrous skin, which is hairless skin. Glabrous skin can be found on the palms and fingertips and the soles of our feet and bottoms of our toes. The anatomy of the skin is depicted in Figure 14.2. Meissner’s corpuscle Free nerve ending Merkel’s discs Pacinian corpuscle Ruf ni ending Hairs FIGURE 14.2 Touch Receptors in a Cross-Section of Skin This illustration shows the dermis, the epidermis, and the different types of nerve endings found in the skin. Epidermis: the outer layer of the skin Dermis: the inner layer of the skin, which also houses touch receptors 404 Sensation and Perception Mechanoreception Touch perception occurs when the skin is moved or touched. This includes indentation, such as when a finger presses against your skin; vibration, such as when you touch an active electric toothbrush; and stretching, such as when someone scratches your back or pulls on your skin. Such mechanical stimulation of the skin activates one or more of the four types of mechanoreceptors in your skin. Mechanoreceptors are sensory receptors in the skin that transduce physical movement on the skin into neural signals, which are sent to the brain. Now, take a deep breath, as we need to get a bit technical and introduce some jargon. There is no avoiding it—touch perception is complex, with multiple kinds of receptors, each with a multisyllabic name. Take your time and repeat the material until you feel comfortable with the terms. The four types of mechanoreceptors are called SAI mechanoreceptors, SAII mech- anoreceptors, FAI mechanoreceptors, and FAII mechanoreceptors. Each mechanore- ceptor differs on a number of relevant dimensions, which are important to producing a complex touch response. The mechanoreceptors are distinguishable anatomically and also in terms of their functional capacities. SA stands for “slow-adapting,” whereas FA stands for “fast-adapting.” The mechanoreceptors labeled with a Roman numeral I have small receptive fields, whereas the II receptors have larger receptive fields. As you read through these distinctions, think about how they map onto other sensory distinc- tions, such as the differences between the magnocellular tract and the parvocellular tract in vision. The slow-adapting mechanoreceptors (SAI and SAII) produce a steady stream of neural response when the skin is deformed, that is, a sustained response that continues for as long as the skin is stimulated. This is important because we want to know if there is continued pressure on our skin, such as when someone is holding your arm. The fast-adapting mechanoreceptors (FAI and FAII) respond vigorously when the skin is first touched and then again when the stimulus ends. This is important because it gives us information about changes on our skin—when something contacts us and when it is no longer in contact with us. However, if the pressure continues, FA mechanoreceptor responses lessen and do not increase until the stimulus is over. Thus, the SA mecha- noreceptors tell us about continued pressure on the skin, such as the constant contact between our skin and our shirts. The FA receptors give us information about the begin- ning and ending of touches and about temporary stimulation of the skin, such as when that annoying mosquito lands on our arm. The SAI and FAI receptors have relatively small receptive fields with densely packed receptors. This gives them high spatial resolution, allowing us to detect exactly where on the skin a stimulation occurs. That is, SAI and FAI fibers help us detect small objects and pinpoint objects on our skin. Again, this is useful for detecting the location of the mos- quito on the skin and for fine manipulation of objects, such as when we thread a needle. Both SAI and FAI mechanoreceptors are close to the skin’s surface. In contrast, the SAII and FAII mechanoreceptors are deeper in the skin. The SAII and FAII receptors have larger receptive fields with more dispersed mechanoreceptors. This means that their spatial resolution is not as good as the SAI and FAI mechanoreceptors, but they have a lower threshold or higher sensitivity to light touch. Thus, they help us feel very light touches, as they sum across space, but they are not as precise at determin- ing where on the skin these touches occur. Thus, when an entire hand brushes against our arm, the SAII and FAII mechanoreceptors let us know that our arm is being moved. These differences are summarized in Table 14.1. You can see a video of the actions of the mechanoreceptors on ISLE 14.1. We now review each mechanoreceptor’s structure and function. Mechanoreceptors: the sensory receptors in the skin that transduce physical movement on the skin into neural signals, which are sent to the brain SAI mechanoreceptors: slow-adapting receptors using Merkel cells, with small receptive fields, densely packed near the surface of the skin SAII mechanoreceptors: slow-adapting receptors using Ruffini endings, with large receptive fields, more widely distributed, deeper in the skin FAI mechanoreceptors: fast-adapting receptors, with Meissner corpuscle endings and small receptive fields, densely packed near the surface of the skin FAII mechanoreceptors: fast-adapting receptors with Pacinian corpuscle endings and large receptive fields, more widely distributed, deeper in the skin Meissner corpuscles: specialized transduction cells in FAI mechanoreceptors Pacinian corpuscles: specialized transduction cells in FAII mechanoreceptors Merkel cells: specialized transduction cells in SAI mechanoreceptors ISLE 14.1 Action of Mechanoreceptors 405 Chapter 14: Touch and Pain SAI Mechanoreceptors SAI mechanoreceptors, using Merkel cells, have a sustained response to continued pressure, giving them maximum response to steady pressure. They have small receptive fields and thus have high spatial sensitivity. They respond best to vibrations at low frequencies. They are important for the touch perceptions of pattern and texture. Because of their high spatial sensitivity, SAI mechanore- ceptors are responsible for two-point thresholds, that is, the minimum distance at which a person can detect two touches instead of just one (Figure 14.3). Thus, these cells are important when we need fine manual control, especially without visual feedback. SAI mechanoreceptors are also critical for blind individuals reading Braille. Two- point threshold responses vary across the skin’s surface. Where it is most sensitive, such as on the fingertips and lips, we find the highest den- sity of SAI mechanoreceptors. Where it is least sensitive, such as on the back and legs, we find the lowest density of SAI mechanoreceptors. For a discussion and illustration of the relation of SAI mechanoreceptors and the Aristotle illusion, see ISLE 14.2. SAII Mechanoreceptors SAII mechanoreceptors, using Ruffini end- ings, have a sustained response to contin- ued pressure, also giving them maximum response to steady pressure. However, they are also good at stretching from side to side, making them crucial for object grasp- ing. Thus, when you pick up your spoon to eat your soup, your SAII mechanoreceptors will be critical. SAII mechanoreceptors have much larger receptive fields than SAI mech- anoreceptors, and so they are more vital for detecting the presence of a light touch than for pinpointing where it occurs. FAI Mechanoreceptors FAI mechanoreceptors respond to the onset and offset of a stimulus, that is, when a stimulus starts and ends. These mecha- noreceptors with Meissner corpuscle end- ings also have small receptive fields for good spatial accuracy. They also respond well to low-frequency vibrations. They are especially good at detecting “slip,” that is, when an object is sliding across the surface of the skin. As such, they are useful in avoiding dropping objects, as they detect an object as it begins to slip away from your hands. These mechanoreceptors are also FIGURE 14.3 A Model of the Body’s Two-Point Threshold Response In this diagram, a person is touched at two points on the skin or one point on the skin. The person must determine if there is one point or two points of touch. Sensitivity to two-point touch is better in parts of the body with higher densities of SAI mechanoreceptors. Two- point thresholds are better on the fingers than on the back. TABLE 14.1 Mechanoreceptors: Types and Response Properties Fast Adapting Slow Adapting Small receptive field Larger receptive field Small receptive field Larger receptive field FAI FAII SAI SAI Meissner corpuscles Pacinian corpuscles Merkel cells Ruffini endings High spatial resolution Low spatial resolution High spatial resolution Low spatial resolution Upper dermis Lower dermis Upper dermis Dermis Medium sensitivity to temperature variation High Low Low Perceiving change Fine texture Patterns Stretch/feedback ISLE 14.2 Mechanoreceptors and Aristotle’s Illusion 406 Sensation and Perception critical in maintaining grip for sports. For example, when a baseball hits a bat, the force causes slippage of the grip on the bat by the batter. FAI mechanoreceptors detect this slip and cause us to tighten our grip on the bat. FAII Mechanoreceptors FAII mechanoreceptors also respond to the onset and offset of a stimulus. They have Pacinian corpuscle endings and larger receptive fields, as they are less densely packed in the skin. They are also more sensitive to higher frequency vibrations. Experientially, when FAII mechanoreceptors are firing, what we feel is more of a buzz than for the other mech- anoreceptors. Because of their high sensitivity to touch, they are helpful in feeling small pressure on the skin, such as when an insect lands on our skin. These mechanoreceptors are also used when we use fine motor control, such as when writing with a pen or tighten- ing a screw with a small screwdriver. The names of the receptive cells in each of these types of mechanoreceptors are given in Table 14.1. In all four types of mechanoreceptors, the mechanical movement of the ending of the receptor induces a neural signal, which is sent first to the spinal column and then to the brain. The nerve fibers from the mechanoreceptors are myelinated, which allows fast transmission of these signals. Interestingly, a new kind of unmyelinated mechanoreceptor has been discovered (Olausson, Wessberg, Morrison, McGlone, & Valbo, 2010). This new fifth mechanoreceptor is known as the C-tactile mechanoreceptor. C-tactile mechanorecep- tors are present only in hairy skin and respond to slow gentle movements on the skin. When stimulated, C-tactile mechanoreceptors induce a pleasant feeling. C-tactile fibers project to the insular cortex, which also receives input from pleasant tastes and odors. TEST YOUR KNOWLEDGE 1. Why is it said that the skin senses are a set of multiple senses rather than a single sense? 2. Describe the difference between slow-adapting and fast-adapting mechanoreceptors. How is each important for perceiving different aspects of the perceptual world? Proprioception: Perceiving Limb Position Proprioception is perhaps our most unique sense, as it is designed to monitor not the external world but the internal world. The mechanoreceptors, the rods and cones of our retinae, and the hair cells of our cochleae in our auditory system are all designed to pick up features external to us, whereas proprioception gives us our awareness of how our own bodies are positioned. Thus, you do not have to see your arms to know that they are in front of your torso, extended outward (doing what you normally do—playing with your phone). You also do not need to see or even feel your feet touching the floor to know that your legs are below you and that you are in a sitting position. We define proprio- ception as the perception of the movements and position of our limbs. You may also see the term kinesthesis, which refers to our awareness or perception of bodily movements. There are three different kinds of sensory receptors in our bodies that provide us with information about limb movement and position. They are the muscle spindles, joint recep- tors, and Golgi tendon organs. We consider each in turn. Muscle spindles are a type of mus- cle cells that have receptors connected to them that sense information about muscle length and therefore muscle action. As muscles contract and lengthen, different neural signals are transduced in the spindles, giving the brain information on muscle length and stretch. Joint receptors are receptors found in each joint that sense information about the angle of the joint. This provides the brain with information about how far the limb is stretched and Ruffini endings: specialized transduction cells in SAII mechanoreceptors Proprioception: the perception of the movements and position of our limbs Muscle spindles: receptors embedded in the muscles that sense information about muscle length and therefore muscle action Joint receptors: receptors found in each joint that sense information about the angle of the joint 407 Chapter 14: Touch and Pain whether it can be stretched any farther before it must be contracted. Golgi tendon organs are receptors in the tendons that measure the force of a muscle’s contraction. Because force will be a function of how many muscle fibers are used, Golgi tendon organs can provide feedback as to how much more force has been applied by that muscle (see Figure 14.4). In Figure 14.4, the afferent fibers are the sensory fibers carrying information to the brain. Because proprioception is directed inward, we seldom consciously focus on it. Rather, we take it for granted that we are aware of the position of our body, how much force we are asserting, and what muscles are moving. However, if we did not have Golgi tendon organs, we might not be able to gauge whether we are applying enough muscle strength to lift weights at the gym or if we are applying too much muscle strength when we intend to be gently patting a friend on the back. We can be thankful for these receptors when we do not break any bones in the fingers of our friends as we shake their hands. When a limb is under physical pressure and blood is temporarily cut off, we may have the experience of an arm or leg “going to sleep” and we cannot feel the location of the limb. If one arm has fallen to sleep, you might need to move your other arm to find it. This feeling is caused by a temporary inability of the proprioception system to detect limb position. Alcohol consumption affects our proprioception system, making it harder for our sensory receptors to give feedback on limb position. It is for this reason that one of the main tests police administer for sobriety is a proprioception test. When a driver is suspected of driving under the influence of alcohol, a field sobriety test is given. One of these tests is to ask the driver to touch her nose with eyes closed (Figure 14.5). Inaccuracy in this test suggests that alcohol con- sumption has compromised the functioning of the proprioception system. TEST YOUR KNOWLEDGE 1. What is meant by proprioception? What deficits would a person have if that person lost his proprioception systems? 2. What are the differences between muscle spindles, joint receptors, and Golgi tendon organs? Thermoreception We are warm-blooded animals, which means that, to survive, we must keep our internal temperature fairly constant. But internal temperature is influenced by the temperature of our environment. If you live in a part of the world that has cold winters, think about Muscle fibers Afferent nerve fiber Tendon Muscle spindle Muscle fiber Afferent nerve fiber Efferent nerve fiber Capsule Capsule Joint Joint receptors Tendon Muscle FIGURE 14.4 Sensory Receptors for Proprioception This anatomical drawing shows where one can find muscle spindles, joint receptors, and Golgi tendon organs. Yellow D og Productions/The Im age Bank/G etty Im ages FIGURE 14.5 Testing for Sobriety The touching the nose test assesses if the proprioception system has been impaired by alcohol consumption. Golgi tendon organs: receptors in the tendons that measure the force of a muscle’s contraction Afferent fibers: neural fibers that carry sensory information to the central nervous system 408 Sensation and Perception stepping outside when you leave your home for school or work. The wind and cold seem to bind to your face, you shiver, and you pull your coat a little tighter around your body. Compare this with the feeling of leaving your air-conditioned hotel on a summer visit to Phoenix, Las Vegas, or some other very hot and very dry climate. Here, it is the heat that stings you. Few people would mistake one experience for the other. Thermoreception is the ability to sense changes in temperature on the skin. This can be passive, as when cold air blows against your face on a winter’s morning, or it can be active, as when we touch another’s forehead to determine if that person has a fever. Obviously, thermo- reception has an important function: Being able to sense temperature is critical for survival. Imagine not being able to sense cold and going out into an Arctic blizzard in a swimsuit. Or imagine not being able to sense how hot it is in the Nevada desert in July and wearing a winter coat. In each case, the lack of thermo- reception could have deadly consequences. We must maintain our internal body temperature, and this is why having a fever is an illness. Thermoreceptors are the sensory recep- tors in the skin that signal information about the temperature as measured on the skin. Thermoreceptors respond to a range of skin temperatures, from 17 °C (63 °F) to 43 °C (109 °F). This is the range of tempera- tures experienced by our skin, not the actual ambient temperature. Thus, the weather may be colder than 17 °C or hotter than 43 °C, but our bodies are designed to regulate the condi- tions in and on our bodies. Skin temperatures above 43 °C and below 17 °C are experienced as pain. In general, our skin maintains a surface temperature between 30 °C and 36 °C (86 °F and 97 °F). This temperature range is sometimes called physiological zero, as you feel neither hot nor cold. At these temperatures, our thermoreceptors are mostly inactive. However, when the temperature goes below 30 °C, our cold fibers become active, and when the temperature rises above 36 °C, our warm fibers become active. If you are inside wearing a thick sweater and sitting by a warm open fire, your skin temperature will increase, and your warm fibers will start firing. If you take off your sweater and go out into the blizzard, your skin will rapidly decrease in temperature, and your cold fibers will start increasing their firing rate. Cold fibers are thermoreceptors that fire in response to colder (30 °C and below) temperatures as measured on the skin. Warm fibers are different thermoreceptors that fire in response to warmer temperatures (above 36 °C) as measured on the skin. At intermediate temperatures, both fibers show a steady firing rate. They differ when the temperature rises above 36 °C or falls below 30 °C (Figure 14.6). Cold and warm fibers also respond when you touch objects that are either colder or warmer than the temperature of your skin. For example, when you remove the ice cream container from the freezer, the surface of the ice cream container will feel cold because the temperature of the container is lower than that of your fingers, causing your cold fibers to fire. When you pick up your hot mug of coffee, the mug feels hot because it is causing warm fibers to fire. Both cold and warm fibers also adapt, somewhat, to new temperatures. Thus, when you first jump into a pool, the water, which is colder than your skin, triggers responses in the cold fibers. But after a couple of minutes, your cold fibers adapt, and the pool no longer feels quite so cold. Interestingly, your warm fibers have a secondary peak in sensitivity when they are exposed to very low temperatures. This is also known as a paradoxical heat experience because it is brought about by contact with extremely cold temperatures. The temperatures Cold �bers 10 0 2 4 6 8 10 20 30 40 50 60 Warm �bers Nerve fiber temperature (°C) Im p u ls e s p e r se co n d FIGURE 14.6 Rate of Firing in Response to Skin Temperature (Guyton, 1991) Graph of firing rates of cold and warm nerve fibers. Thermoreception: the ability to sense changes in temperature on the skin Thermoreceptors: the sensory receptors in the skin that signal information about the temperature as measured on the skin Cold fibers: thermoreceptors that fire in response to colder (30 °C and below) temperatures as measured on the skin Warm fibers: thermoreceptors that fire in response to warmer temperatures (above 36 °C) as measured on the skin Pain: the perception and unpleasant experience of actual or threatened tissue damage 409 Chapter 14: Touch and Pain are so low that they also induce pain responses. Nonetheless, some of you may have expe- rienced this during minor medical treatment. If you have ever had a skin abnormality, such as a wart, treated with liquid nitrogen, you know the experience of paradoxical heat. Liquid nitrogen is kept at about –200 °C (–328 °F). When it is applied to our skin, however, we feel pain and heat, not coldness, because of this secondary peak for the warm fibers. Some Antarctic explorers have also described feeling paradoxical heat when being buffeted by 160 km/h winds in –60 °C cold. See ISLE 14.3 for instructions on a safe way to demonstrate this phenomenon to yourself. TEST YOUR KNOWLEDGE 1. What is thermoreception? 2. Why is it likely that animals, such as ourselves, evolved to have two separate systems, one for detecting cold and one for detecting heat? What advantages does this have over a single-receptor system that simply detects temperature, much as a thermometer might? Nociception and the Perception of Pain Pain is the perception and the unpleasant experience of actual or threatened tissue damage (Figure 14.7). Thus, pain is the result of activation of receptors in our skin and elsewhere, as well as the unpleasant subjective feeling asso- ciated with it. As we will see in this chapter, pain can arise from several different causes and may be accompanied by emotional distress that may act to increase the experience of pain. In this section, we focus on the sensory receptors in the skin that transduce a neural signal, causing an experi- ence of pain in the affected area. This type of pain is called nociceptive pain. Nociceptive pain is the pain that devel- ops from tissue damage that causes nociceptors in the skin to fire. This type of pain occurs from direct trauma to the skin, such as from cutting, puncturing, pinching, heating, freezing, or coming in con- tact with toxic chemicals (ouch!). These types of events cause nociceptors in the skin to fire, leading to a signal to the brain. Nociceptors are sensory receptors in the skin that, when activated, cause us to feel pain. They are found in both the epidermis and dermis. Although we focus on the nociceptors in the skin, there are nociceptors in many other areas of the body, including muscle, bone, and the digestive system. There are also other forms of pain, such as inflammatory pain, in which an inflamed or swollen region causes changes in nearby nociceptors. Neuropathic pain occurs when the nervous system itself is damaged. Nociceptors are anatomically distinguishable from other skin recep- tors. Nociceptors are often referred to as free nerve endings because of their anatomical appearance (Figure 14.8). Nociceptors are divided into two main types: A-delta fibers and C-fibers. A-delta fibers are myelinated, so they conduct signals rapidly and respond to both heat and pressure, both of which can induce pain. C-fibers are nonmyelinated and hence much slower; they respond to pressure, extreme degrees of either heat or cold, and toxic chemicals. The signals from both of these free nerve endings are experienced as pain. A-delta fibers are associated with the stinging feeling of pain. This stinging pain is often what we experience first when we are injured. C-fibers are associated with the more chronic experience of throbbing pain (Williams & Purves, 2001). This pain may be a ISLE 14.3 Heat Grille FIGURE 14.7 Pain Is Unpleasant We experience pain as a result of threatened or actual tissue damage. Corey Rich/A urora/G etty Im ages Nociceptive pain: pain that develops from tissue damage that causes nociceptors in the skin to fire Nociceptors: sensory receptors in the skin that, when activated, cause us to feel pain; they are found in both the epidermis and the dermis A-delta fibers: myelinated nociceptors that conduct signals rapidly and respond to both heat and pressure C-fibers: nonmyelinated nociceptors that are slower and respond to pressure, extreme degrees of either heat or cold, and toxic chemicals 410 Sensation and Perception bit delayed from the actual accident. Think about touching a hot stove top. You first get a searing pain, and you quickly move your hand away from the hot surface. This sear- ing pain is the result of the fast action of the A-delta fibers. After you start treating your injury, you start to feel continued dull throb- bing in your finger, which is the result of the C-fiber response (see Figure 14.8 for a figure showing the nociceptors). It probably has crossed every human being’s mind at one time or another how much better life would be if we did not expe- rience pain at all. Certainly, anyone who has experienced serious pain can identify with the wish not to have pain at all. Something as seemingly simple as an infected tooth can cause excruciating and debilitating pain, as anyone who has ever gone through a root canal can tell you. We certainly pay a great deal of money as a society to reduce and control pain. However, the argument can be made that pain serves an important evolu- tionary function, by alerting an organism to tissue damage, allowing it to take steps to minimize or repair the tissue damage or move away from the source of that damage. Through evolution, pain evolved pre- sumably because organisms that moved away from sources of pain were more likely to survive than those that did not. When we do not experience pain, we might not act to avoid these consequences. Indeed, Melzack and Wall (1988) described a case of a human patient lacking in nociceptors. This woman experienced numerous avoidable injuries because she lacked pain perception. These injuries included burns from not removing her hand from a stove top, biting her own tongue, early joint problems, and, ultimately, death from curable infections. TEST YOUR KNOWLEDGE 1. Describe the differences between A-delta fibers and C-fibers. How does each contribute to the perception of pain? 2. Develop a diagram that relates different receptors to different sensations arising from the skin. NEURAL PATHWAYS 14.2 Sketch the two main pathways from receptors in the skin to the somatosensory cortex in the brain. In May, off the coast of the Dominican Republic, boats can take you out to go free div- ing with the humpback whales that migrate close to the shores of that country. Diving with some magnificent creatures is an amazing experience, as humans are dwarfed by Pacinian corpusclePacPacPacPacPacPacPacPacPacPacacPacPacPacPacaPacPacPacPacPacPacPacPacacPacPacccPacPPacPaPacPacPacPacacPacPPPacPacPPacPaPPacPPaccPPacPPaPaPacaaPacPacPaPacccacccciniiniiniinnnniiniiniiniiniiniiniiniiniiniinininininiiininininnnniniiniinnninnnninniii iiiniiniinnnnniniinnninnnniinnnnnnianaan ananananannanannananananaaaaaanan anaaaaaaaanananananaaananaaaannnan an anananaaannnnananananaan nnnnannanaanann ananannanan aanaan annnannnaaaaannannnaaannnan aaaaaaanaaannannn anaaannnnaaannannaan aannannaaan nnnnn anannnnnannnnaaa corcorcorcorcorcorcorcorcorcccorcococorcorcococorcoroocorcorrorcororcorcocorccccocoococorcororcorcorcorcorcorcorcorccorccoccooocorcorrcorcorcorcorcorcorccorcocccoorrcorcocorccorcocccorcororrcocorccocorrrcoccccccocorrorcococcocccocorcoooorrcorcoccccooroorccccc rcccccocoorcccccocororccccoooorcccccccoooorcccoroooooorcccoooccocooooo puspuspuspupppuspuspupuspuspuspupuspuspuspuuuupuspuspuspuspuspuspuspuspuspuspuspuspuupuspuspupupupuuupuuuspuupuspuspuppususpuupuspuspusppupupuspuspuspuspuspuupusppupuppuupuupuspppppupuuupussusppppppppuuuupupuuuuppuuuupppp cleclecleeclecleclecleclecleclecleclecleclecleclelecleclecleclcleclecleclecleecleclecleclecleleeecleleeeeecleclleleeecleleeelcleclecleclllclleccleeecleleecc Nociceptors FIGURE 14.8 Nociceptors in the Skin Nociceptors are located just below the epidermis, in the dermis. Nociceptors are activated when tissue is damaged, leading to the experience of pain. Dorsal root ganglion: a node on the spine where one finds nerve cells carrying signals from sensory organs toward the somatosensory areas of the brain Dorsal root: the end of the spinal nerve where sensory information enters the spinal cord Ventral root: the end of the spinal cord where motor information leaves the spinal cord 411 Chapter 14: Touch and Pain the incredible size of these whales, which can exceed 16 m (52 ft). You can swim right up to these creatures, which do not see divers as a threat and are sometimes curious about their presence. You are not supposed to touch the animals at all, but what would happen if you brushed against the tail of a whale (Figure 14.9)? Given the distance between the whale’s tail and the whale’s brain, it would take upward of a second for the whale to feel your touch. That is, if instead of a gentle brush, you pricked the tail with a pin, it would be nearly a second before the whale experienced pain (Hof & van der Gucht, 2007). Of course, reflex circuits would cause the whale to swish its tail away from you much ear- lier, hopefully filling your mouth and mask with seawater (after all, you pricked the whale’s tail). In humans, the distance nerve fibers must travel is much less than in whales, but it is still much farther for touch and pain than it is for sensory systems already arranged in the head (vision, audition). For a tall person, it may still be 2 m (6 ft 6 in.) from toes to somatosensory cortex. Thus, when you stub your toe, the neural message has a greater distance to travel than when someone yanks your ear or when light hits your retinae. As a function of this distributed feature of the somatosensory system, we see some major differences, too, in the pathways for these senses arising from the skin and body. All right, now, get your attention in gear, as we start with the complex anatomy involved in transmitting the neural signals for touch, temperature, and pain from the skin to the brain. We will also consider the areas of the brain responsible for the touch senses and pain. Starting with the neural pathways, consider a nerve ending in the skin of the palm of your hand. This nerve end- ing sends its axon into a nerve bundle, where it is joined by many other axons from adjacent nerve fibers. These neurons are oddly constructed. The part of the cell that has the nucleus is not at the beginning of the neuron but just before it enters the spinal cord, in what is called the dorsal root ganglion. The cells enter the spinal cord in what is called the dorsal root (the ventral root transmits information to the muscles). The dorsal root forms one of the two branches of nerves that enter and leave the spinal cord (Figure 14.10). The axons of the sensory cells that pass through the dorsal root enter the dorsal (i.e., back) part of the spinal column. (You can see an interactive model of how our touch and pain senses enter the spinal cord in ISLE 14.4.) Just to help you keep track of directions, FIGURE 14.9 A Whale’s Tail A humpback whale’s tail may be as far as 16 m from its brain. Thus, the neural signal that occurs when the diver touches the tail may take nearly a second to reach the somatosensory cortex of the whale. © iStockphoto.com /ShaneG ross ISLE 14.4 Somatosensory Pathways FIGURE 14.10 The Dorsal Root Ganglion Nerve fibers from nociceptors enter the dorsal root ganglion. In this illustration, you can see the anatomy of the ganglion and its exit route into the spinal column. Sensory neuron to brain Dorsal root Dorsal root ganglion Sensory neuron from hand Spinal neuron Biceps muscle Ventral root Motor neuron Interneuron Ventral horn Motor neuron from brain 412 Sensation and Perception dorsal refers to back and ventral refers to front, when we are talking about our body as a whole. Once in the spinal column, this somato- sensory information is divided into two distinct tracts or pathways, which travel up the spinal column to the brain. It is important to keep these two tracts distinct, as they serve different functions, and carry different sensations. Both the anatomy and the terminology are tricky. So, stay with the text and review the text and figures often. You can also follow along on ISLE 14.4 to help you learn about these different pathways. One tract is called the dorsal column–medial lemniscal pathway; this pathway carries information from the mechanoreceptors (i.e., tactile perception) and from the proprioceptors (i.e., muscle position perception). The other track is called the spinothalamic pathway, and it carries information from the nociceptors (pain) and the thermoreceptors (tempera- ture). The names are a mouthful, but the anatomy is tricky too. The dorsal column– medial lemniscal pathway travels on the dorsal (remember, the back) side of the spi- nal column and on the ipsilateral side of input from the skin (the same side as the input from the body). The dorsal column– medial lemniscal pathway makes a synapse in the medulla of the brain, where it crosses over to the contralateral side (the opposite side of the input from the body). It then ascends into the brain as the medial lemnis- cus. From the medulla, it travels to the ventral posterior nucleus of the thalamus and from there to the somatosensory cortex (see Figure 14.11 for a graphical display of this tract). The spinothalamic pathway’s fibers first synapse in the spinal cord and then cross over to the contralateral side within the spinal column and then ascend toward the brain. That is, information from the left side of the body travels up the right side of the spinal cord in the right spinothalamic pathway, and information from the right side of the body travels up the left spinothalamic pathway. Then, without any additional synapses, the spinothalamic pathway goes directly to the ventral posterior nucleus of the thalamus and from there to the somatosensory cortex. This anatomy is depicted in Figure 14.11 and ISLE 14.4. The goal of these pathways is to get information to the brain for processing. There are also reflex channels in the spinal cord that take the information and send it right back to the muscles to allow fast reactions to dangerous situations. To experience the feeling, though, the information must be sent to the somatosensory cortex. You may have experienced this at one time or another. If you ever accidentally (or on purpose) placed your hand on a hot stove top, you may have noticed that you jerked your hand away before you actually felt the acute pain. Like the whale’s tail, it takes a bit longer Anterior cingulate cortex Signals to anterior cingulate cortex Signals to insular cortex Somatosensory cortex Spinothalamic tract Medulla Spinal cord Ventral posterior nucleus of thalamus Medial lemniscus First synapse (dorsal) First synapse (spinothalamic) Dorsal root ganglion Dorsal horn Thalamus Medulla Spinal cord Somatosensory cortex Nerve fibers carrying neural signals for tactile perception and proprioception Nerve fibers carrying neural signals for nociception and thermoreception Spinothalamic pathway Dorsal column–medial lemniscal pathway FIGURE 14.11 The Two Pathways to the Brain This anatomical diagram shows the dorsal column–medial lemniscal pathway and the spinothalamic pathway. Dorsal: in or toward the back of the body; in the head, it means at the top or toward the top Ventral: in or toward the front of the body; in the head, it means at the bottom or toward the bottom Dorsal column–medial lemniscal pathway: a pathway for the mechanoreceptors (tactile perception) and proprioceptors (muscle position) that travels up the spinal column on the ipsilateral side and crosses to the contralateral side in the medulla Ipsilateral: literally, same (ipsi) side (lateral), meaning, in this context, that sensory information is on the same side of the nervous system as it entered 413 Chapter 14: Touch and Pain for the information to get to your brain for you to experience pain than it does for the reflex arc to allow you to remove your hand quickly. Somatosensory Cortex The somatosensory cortex is an area in the parietal lobe of the cerebral cortex devoted to processing the information coming from the skin senses. The somatosensory cortex is in the very anterior (front) of the parietal lobe, just behind the central sulcus, which divides the parietal lobe from the frontal lobe. The most anterior area is called the primary somatosensory cortex (S1), and just posterior to it is the secondary somatosensory cortex (S2) (Figure 14.12). The primary somatosensory cortex receives input from the ventral posterior nucleus of the thalamus and thus receives input from both the dorsal column–medial lemniscal pathway and the spinothalamic pathway. As we will discuss in more detail, different parts of the somatosensory cortex process information from different parts of the body. Directly across the central sulcus lies an area in the frontal lobe devoted to the con- trol of movement. The movement area in the frontal lobe maps mostly onto the primary somatosensory cortex. This means that if the primary somatosensory cortex receives input from the left thumb, the area adjacent to it in the frontal lobe motor cortex will be involved in move- ment of this same thumb. A brief historical interlude is called for here. Much of what we know of the function of the somatosensory cortex comes from pioneering studies by North American neurosurgeon Wilder Penfield (1891–1976). Penfield was interested in surgical ways to relieve the agony of epilepsy. He reasoned that if damaged brain tissue was surgi- cally removed, patients with severe epilepsy might improve. However, it was important to know the function of the brain areas in question before they were surgically removed. Therefore, Penfield pio- neered a technique of directly stimulating areas of the brain with an electrical probe and observ- ing the behavior of the patient. This required only local anesthetic to the skull and skin, as the brain itself has no nociceptors. The brain feels no pain when it is stimulated, just the pain from the body traveling into the head to be experienced. When probing along the primary somatosensory cor- tex, Penfield found that if he stimulated one area, the patient felt a tingling in a specific area of the body. Moving to an adjacent area of the somato- sensory cortex, the tingling would occur in an adjacent area of skin. In this way, Penfield was able to map out the relation between the primary somatosensory cortex and the skin’s surface (Penfield & Rasmussen, 1950). You can see this map in Figure 14.13. S1 S1 S2 Postcentral sulcus Postcentral sulcus 1 2 3b 3a Posterior parietal cortex Central sulcus Central sulcus Lateral sulcus FIGURE 14.12 Somatosensory Cortex Regions S1 and S2 are important in somatosensory perception, from touch to pain. Shoulder Hip Knee Toes Trunk Neck Arm Elbow Wrist Hand Fingers Thumb Eye Brow Face Lips Swallowing Jaw Tongue Hip nee es b row Face LipL s Swallowing Jaw FIGURE 14.13 Somatosensory Cortex A homunculus drawing of the somatosensory cortex as conceived by Penfield and Rasmussen (1950). Note the larger areas in the somatosensory cortex devoted to the fingers and mouth and the relatively smaller areas devoted to the back and the legs. 414 Sensation and Perception The primary somatosensory cortex maintains a somato- topic map of the body. Recall that soma means “body,” and therefore a somatotopic map is a map of the body. In the somatotopic map, the skin of the body maps onto the sur- face of the primary somatosensory cortex in a systematic way, just as we saw with the mapping of the retina in the retinotopic map and the mapping of the basilar membrane in the tonotopic map. When you scratch your chin, this stimulates an area in the primary somatosensory cortex responsible for perceiving sensations on your chin. Scratch the tip of your nose, and an area slightly higher in the somatotopic map will be stimulated. Rub your forehead, and an adjacent area to that in the somatosensory cortex becomes active. In fact, as Figure 14.14 demonstrates, the somatosensory cortex actually holds multiple maps of the skin surface. But throughout these maps, you can see the feature Penfield originally noted. The maps are distorted. Areas of the body for which we need greater sensitivity (e.g., hands, mouth) have a greater representation in the somatosensory cortex maps (i.e., they take more area on the map) than do areas of the body for which we do not need as high sensitivity (e.g., back, legs). It is for this reason that the homunculus drawing (Figure 14.13) looks so odd. Penfield and Rasmussen drew an image of a human enlarging those areas that receive greater space in the somatosensory cor- tex and shrinking those areas that receive less space in the somatosensory cortex. This figure is drawn over the top of the somatosensory cortex in Figure 14.13. Suborganization of the Somatosensory Cortex The primary somatosensory cortex (S1) is divided into three distinct neuroanatomical regions. One of these regions is then divided again into two regions. These regions are called Area 1, Area 2, Area 3A, and Area 3B. The terminology is a bit confusing as these areas—1, 2, 3A, and 3B—are all part of S1, the primary somatosensory cortex. Each area has a distinct map of the body’s surface. In addition, each area receives input from a particular tract in the somatosensory system. Area 1 receives input from mech- anoreceptors and is therefore the primary area for tactile perception. Area 2 receives input from proprioceptors in the muscles and tendons (as does Area 3A). Area 3A receives input from nociceptors. Area 3B also receives input from nociceptors, along with input from the mechanoreceptors. Thus, each subarea of the primary somatosen- sory cortex processes different fundamental sensations that arise from the skin and body. These areas then send information to S2 (the secondary somatosensory cortex), which is the somatosensory system’s “what” channel for identifying the nature of the touched object (Hsiao, 2008). That is, this channel, sometimes called the ventral sys- tem, is critical for using touch information to identify both the shape of an object and the identity of that object. Thus, this area allows us to discriminate softballs from grapefruits (Figure 14.15). There is also a “where” channel or dorsal system in the somatosensory system. The dorsal system allows us to control guided movements based on input from the somato- sensory system. For example, we use feedback from this system to adjust our grip on Leg Foot Toes Genitalia Throat Tongue Teeth, jaw, gums Trunk S1 S2 Neck Head Shoulder Primary somatosensory cortex (S1) Area 1 Area 2 Area 3 Secondary somatosensory cortex (S2) Arm Elbow Forearm Hand Digit 5 Digit 4 Digit 3 Digit 2 Thumb Eyes Nose Face Upper lip Lower lip Chin FIGURE 14.14 The Somatotopic Map A more modern view of the somatotopic map, showing multiple regions of the somatosensory cortex and both S1 and S2. Contralateral: literally, opposite (contra) side (lateral), meaning, in this context, that sensory information is on the side of the nervous system opposite the one from which it entered Somatosensory cortex: an area in the parietal lobe of the cerebral cortex devoted to processing the information coming from the skin senses Ventral posterior nucleus of the thalamus: an area in the thalamus that receives input from both the dorsal column– medial lemniscal pathway and the spinothalamic pathway Spinothalamic pathway: a pathway for the nociceptors (pain) and thermoreceptors (temperature) that travels up the contralateral side of the spinal column; does not synapse in the brain until the ventral posterior nucleus of the thalamus 415 Chapter 14: Touch and Pain a large textbook when it starts slipping away from our hands. When we feel that the barbell we are about to lift is heavy, feedback from the somatosensory system tightens our grip. The “where” channel continues from areas in S1 to the posterior, or back part of, the parietal cor- tex, an area involved in the control of action, and then back forward to the premotor cortex in the frontal lobe (Hsiao, 2008). Left out of the previous description is the pathway for temperature information. It turns out that temperature information takes a slightly different pathway when it makes its way to the cortex. Thermoreceptors send axons up the spinothalamic pathway that travel to the ventral posterior nucleus of the thalamus and from there to S1. In con- trast to other somatosensory systems, temperature information then goes to areas of the frontal lobe, including the insular cortex and the anterior cingulate cortex (Hua, Strigo, Baxter, Johnson, & Craig, 2005). The insular cortex and anterior cingulate cortex have complex roles in cognition and are not typically associated with sensory perception. Thus, their roles in perceiving temperature remain to be determined. But given that both areas are also associated with emotion, it may explain why some peo- ple feel so passionate about the weather. Pathways for Pain When we touch a hot stove, or prick a finger with a sewing needle, or feel the injection of Novocain into our gums at the dentist’s office, we feel pain. Much to our regret, nothing could be more real than the uncomfortable, distracting, and unpleasant experience of pain. We may try to avoid it, we may dread it, we may develop ways of coping with it, but from time to time, human beings experience pain, almost invariably unpleasant. However, it turns out that the experience of pain arises from interactions among multiple systems, not just the nociceptors. Pain can also be modified by emotion and cognition, which act to change the pain experience in very real ways. For example, we describe in the next chap- ter how the experience of pain can enhance the perception of flavor in spicy foods, a rare situation in which pain is perceived as a positive. The neuroscience of pain is therefore a fascinating story of the complex interactions of bottom-up (nociceptive) and top-down (cognitive and emotional) processes. As we discussed earlier, nociceptors in the skin detect damage or trauma at the skin and transmit that information up the spinothalamic pathway. However, there is also a downward pathway, leading from higher centers in the brain and heading down into the spinal column. The information coming from the brain can inhibit the flow of infor- mation upward toward the brain from the nociceptors. When it does so, the experience of pain is reduced or even inhibited. When the downward signals are excitatory, the transmission of pain information is blocked or inhibited. This view of pain perception has been called the gate control theory (Melzack & Wall, 1988). It is depicted in Figure 14.16 and can be seen in ISLE 14.5. We also know where gate control occurs. The nociceptors first synapse in the spinal cord in an area called the substantia gelatinosa of the dorsal horn. It is here that neural signals from the brain can inhibit the upward flow of pain information. Somatotopic map: a feature whereby the skin of the body maps onto the surface of the primary somatosensory cortex in a systematic way Homunculus: a drawing of a human in which the proportions of the body parts match the relative sizes each body part has on the somatotopic map. Gate control theory: a model that allows for top-down control of the pain signal coming up the spinal cord Substantia gelatinosa: the region in the dorsal horn where neurons meet Dorsal horn: an area of the spinal cord that receives input from nociceptors and feedback from the brain FIGURE 14.15 Shape and Texture The “what” channel allows us to discriminate and identify objects that have similar shapes and sizes, such as this softball and this grapefruit. © iStockphoto.com /JoeG ough © iS to ck ph ot o. co m /m ic ha kl oo tw ijk ISLE 14.5 Melzack and Wall’s Gate Control Theory 416 Sensation and Perception Think of the implications of the gate control model. Think, if you will for just a moment, of a prisoner being tortured. The fear a prisoner may experience in this cir- cumstance may ensure that the gate control is wide open, and all pain signals are head- ing upstream unchecked. Indeed, a small pin- prick may cause enormous pain under these circumstances. Contrast that with a visit to the dentist. The dentist is not your enemy but your doctor. You know your dentist is not out to maim you but rather to make sure your teeth are healthy. The comfort of that knowledge may be enough to activate the gate control mechanism and effectively lower the experience of pain. Visits to the dentist may still involve pain, but because we know the reasons behind this pain, our gate control mechanism may mute the pain. This leads us to the role of the anterior cingulate cortex in emotional responses to pain (Davis, Taylor, Crawley, Wood, & Mikulis, 1997). The anterior cingulate cor- tex receives input from the primary and secondary somatosensory cortices (S1 and S2) and responds to pain caused by pinches, pricks, and extreme heat and cold. Indeed, in one study, patients undergoing neurosur- gery showed strong activity in their ante- rior cingulate cortices when experiencing pain (Hutchinson, Davis, Lozano, Tasker, & Dostrovsky, 1999). In another study (Rainville, Duncan, Price, Carrier, & Bushnell, 1997), participants were hypnotized and then had their hands placed in either painfully hot water or tepid water. S1 and S2 were activated regardless of hypnotic suggestions given to participants, but the activity in the anterior cingulate cortex varied as a function of hypnosis condition. When participants were led to believe that more pain was coming, more activity was seen in the anterior cingulate cortex than when people were led to believe that pain was diminishing. Another study has shown that people show activity in the anterior cingulate cortex when other people are experiencing pain (Singer et al., 2004). In this way, the anterior cingulate cortex is seen as an area important to the emotional component of pain perception. It can regulate pain and likely can initiate the gate control mechanism to inhibit pain. TEST YOUR KNOWLEDGE 1. Diagram the dorsal column–medial lemniscal pathway and the spinothalamic pathways. Where is each located, and what somatosensory systems is each responsible for? 2. Explain to your roommate the significance of the somatosensory homunculus in terms of what it tells us about sensitivity to different parts of the body. The brain’s response to painful stimuli releases endorphins in the periaqueductal gray area, which sends an inhibitory message to the spinal cord Midbrain Periaqueductal gray area Pons Medulla Spinal cord Enkephalin Message from pain receptors Substance P Reduced pain message to brain FIGURE 14.16 Gate Control Theory This model shows both how pain arises in the nociceptors and how it is transmitted to the brain. It also depicts top-down control of the pain signal coming down from the brain through the spinal column. Anterior cingulate cortex: a region in the prefrontal lobe of the brain associated with the emotional experience of unpleasantness during pain perception 417 Chapter 14: Touch and Pain THE NEUROCHEMISTRY OF PAIN: ENDOGENOUS OPIOIDS 14.3 Examine the role of endogenous opioids in controlling pain perception. As we have described, pain is useful because it alerts an organ- ism to the potential of trauma to the body. This is useful because it allows the organism to move away from or reduce the poten- tial for further injury. We learn from pain. However, there may be some situations in which an organism must continue to put itself in harm’s way. For example, think of the pain an ante- lope may feel from a thorn in its foot as it runs away from a cheetah. In other circumstances, this pain might be a signal to slow down and see if the antelope can shake out the thorn, but with the cheetah in pursuit, the pain must be ignored because a greater danger needs to be confronted. Similarly, think of a human mountain climber battling the steep slope she must descend to return from the mountain and the pain she is experiencing from a twisted ankle. Giving in to the pain of the ankle means death, so the climber must ignore the pain to continue her descent (Figure 14.17). Humans may also mimic some of these situations in athletic contests. Consider the marathon runner, whose muscles are exhausted and overheated and whose nociceptors are firing at high rates by the middle of a race, even in the best of conditioned athletes. To complete the 26.2 miles, the marathon runner must continue to run despite considerable pain. By now it should not surprise you to learn that the nervous system has mechanisms to reduce pain perception in these circumstances. The brain can release chemicals known as endogenous opioids at various sites in the nervous system, which act as analgesics to reduce pain. Endogenous opioids are neurotransmitters produced by the body that reduce pain, via analgesia, throughout the body. They are called endogenous opioids because opiate drugs such as morphine and heroin have their effects by mimicking or acting like these substances in the body (Froehlich, 1997; Terenius, 2000). Thus, the marathon runner will begin to feel the “runner’s high” when these opioids kick in. The pain associated with muscle fatigue seems to lessen, and the runner can even speed up at the finish. Stories abound of lost hikers who overcame pain and injury to make it to safety. Endogenous opioids may have contributed to many of these self-rescues. Looking closer at the opiate drugs can give us some insight into the operation of these endogenous opioids. Opiates include many drugs, both legal and illegal, that people take externally to reduce pain. Opiates include medically used painkillers, such as codeine, and illegal drugs, such as heroin. Endorphins are a class of endogenous opioids produced by the body, which inhibit pain responses in the central nervous system and thus reduce pain perception. Because artificial opiates mimic the effects of these natural painkillers, they can be effective at reducing pain, although in many cases, the risk for addiction is high as well. For this reason, the use of opiates as painkillers is by prescription only. Even with this precaution, there is a severe epidemic of prescription opiate abuse in this country (Compton, Jones, & Baldwin, 2016). TEST YOUR KNOWLEDGE 1. Explain endogenous opiates. How do they work to control pain? 2. Compare and contrast endogenous opiates and opiates on their control of pain. A scent Xm edia/Photographer’s Choice/G etty Im ages FIGURE 14.17 Control of Pain Perception The climber might feel pain at this point, but the climber must ignore the pain. Letting go at this point could mean death. Endogenous opioids: chemicals produced by the body that reduce pain throughout the body Analgesia: processes that act to reduce pain perception 418 Sensation and Perception THE PERCEPTION OF ITCH 14.4 Contrast the difference between pain and itch and how they interact. Itchiness is usually a minor inconvenience. Dry skin feels itchy, and you give it a quick scratch and move on. A single mosquito bite annoys you with its itchiness, but it is easily ignored. But itchiness can also be tremendously debilitating. Jellyfish stings can itch all over one’s skin with an incredible intensity. This itchiness can lead to self-induced wounds that may get infected. Itchiness can also disrupt sleep patterns and distract attention from necessary tasks. There is a disease called river blindness (technically, onchocercia- sis), which is a major problem in tropical regions of Africa. The disease may affect tissue in the eye, causing partial or complete blindness, but it also causes a terrible itching sensation all over the body. People with river blindness have been known to throw themselves into water, even when crocodiles may be swimming about in that water, just to relieve the itching. If you think about it subjectively, itch is different from pain, even though both may be considered negative sensory experiences. Itchiness causes us to want to scratch the affected skin, a response not usually caused by pain (Figure 14.18). Pain induces a with- drawal response—we move away from pain. Thus, behaviorally, they have different consequences as well. Other animals use itch perception in other ways. Think of your cat and her notorious itchy ears. Cats’ faces and ears are loaded with itch receptors to induce them to rub against landmarks and leave their scents for other cats to detect. Other animals may feel itchy for similar reasons. Recent research shows that itch perception is caused by a different class of recep- tors in our skin. These receptors have been labeled pruriceptors (pruri is a Latin root meaning “itch”). Pruriceptors respond mostly to chemical irritants on the skin rather than tissue damage (LaMotte, Shimada, Green, & Zelterman, 2009). Like nociceptors, pruriceptors also send their axons up the spinothalamic tract, where they eventually synapse in the somatosensory cortex. Pruriceptors have been less well studied than other forms of skin perception, but there is now strong evidence that pruriceptors are anatomically different than nociceptors. The function of the relation of itching and scratching has undergone some scrutiny. One hypothesis is that we scratch to remove the possible irritant that may be bothering us. Thus, a landing mosquito causes our skin to itch, and our scratch removes the insect from our skin. Another hypothesis is that itching causes us to scratch, which then induces actual, albeit minor, tissue damage. That is, the goal of the itchy sensation is to induce us to cause minor tissue damage. This tissue damage initiates an autoimmune response, which acts to reduce the risk caused by the original irritant (LaMotte et al., 2009). Consistent with this view is the finding that pain inhibits itch. That is, when you scratch an irritant on your skin, it may induce mild pain, which causes the itch to decrease as well as possibly an immune response (Ward, Wright, & McMahon, 1996). Thus, for most people, having one’s back scratched feels good at first, but if the scratching is too hard, pain perception takes over. So, the next time you get a mosquito bite, you can feel a bit more comfortable scratching your itch, knowing that in small doses, it may help your body react to the irritant. TEST YOUR KNOWLEDGE 1. Describe the anatomical and functional differences between pain and itch. 2. Evaluate the evolutionary purpose of itch. FIGURE 14.18 The Perception of Itch Skin irritations cause itching, and itching induces us to scratch. Sh ut te rs to ck .c om /n am tip St ud io Pruriceptors: receptors in our skin that respond to mild irritants by producing itch sensations 419 Chapter 14: Touch and Pain HAPTIC PERCEPTION 14.5 Assess the role movement plays in enhancing the perception of touch, allowing us to perceive objects. The retina is a fixed shape, and the basilar membrane waits for sounds to enter the ear to respond and move. But our skin is different. It moves as we move and changes shape as we interact with our environment. Think about find- ing your way around in a dark room. You reach out and touch objects to determine what they are, but you do not just touch them—you pick them up and put your hands around them. Indeed, the determination of shape and identity from touch involves a dynamic process of percep- tion, motor control, and feedback. We cannot do this on ISLE, but try having a friend blindfold you and then give you common objects to identify by touch. When given a small rectangular wafer-shaped metallic object, you feel it, you move it around, and you see if it has moving parts before you confidently declare that it is a flash drive. This dynamic process of object identification by touch is called haptic perception. Haptic perception refers to the active use of touch to identify objects. Haptic perception includes integrating information from mechanoreceptors, proprio- ceptors, thermoreceptors, and perhaps nociceptors as well. Normally, we are not blindfolded, but haptic percep- tion is important in many real-world situations as well. For example, you may need to find a quarter (25-cent piece) in your pocket. You do not need to take out every coin and examine them visually to determine the right coin. Rather, you feel the shapes of the coins in your pocket and use your haptic perception to determine the correct coin. People who work with tools all the time can pick up the desired tool by feel alone. Cooks know when their batter is ready to go in the oven not because it looks right but because they feel it with their hands and know when the texture is correct. And perhaps most important, surgeons may often be required to perform operations by touch, as that information is more readily available than visual information. Thus, haptic percep- tion is relevant in many different situations. Think again about identifying objects while blindfolded. Experiments have examined exactly that ability in people. For example, Klatzky (see Lederman & Klatzky, 2009) gave participants 100 common objects to touch and identify. These objects included eyeglasses, keys, spoons, and coffee cups. Not only were the participants highly accu- rate (nearly 100%), but they made their accurate identifications very quickly. Klatzky, Loomis, Lederman, Wake, and Fujita (1993) showed that wearing a glove while attempt- ing to identify objects interfered with the haptic perception and lowered accuracy. They hypothesized that the glove prevented the receptors from picking up vital information from the objects, thus reducing the accuracy of the judgments. Thus, we can identify objects via touch, and factors that interfere with efficient touch can interfere with object identification. To identify objects, participants make a number of discrete movements, such as lateral motion, applying pressure, following the edges, supporting the object to deter- mine approximate weight, and several others (Figure 14.19). These motions are called Lateral Motion (Texture) Unsupported Holding (Weight) Pressure (Hardness) Enclosure (Global Shape) (Volume) Static Contact (Temperature) Contour Following (Global Shape) (Exact Shape) FIGURE 14.19 Exploratory Procedures (Lederman & Klatzky, 2009) We use our hands to explore objects and determine their shapes and identities. Haptic perception: the active use of touch to identify objects 420 Sensation and Perception exploratory procedures. Exploratory procedures are hand movements we make to identify an object. Thus, for example, we can determine the texture of an object by lateral motion, or moving our hands along its surface. A smooth object will result in a very different pattern on our mechanoreceptors than will a bumpy object. You can tell the difference between your cat and your dog without looking at them by feeling the differences in the shape of their heads. Think about moving your fingers across a rough surface, perhaps sandpaper. As you move your fingers across the sandpaper, the pattern registers in unique ways on the mechanoreceptors of your fingers. And depending on the grain of the sandpaper, different patterns of high and low activity will occur in the FA and SA receptors. This pattern allows the somatosensory cortex to develop a spatial map of what the surface is. This conjecture was put to the experimental test in a clever experiment conducted by Hollins, Bensmaia, and Washburn (2001), who compared the responses of mechanoreceptor fibers across different surfaces. Hollins et al. (2001) did this by inducing a tactile illusion. They started with the assumption that FAII (fast-adapting, large receptive fields) mechanoreceptors would be important in determining the identities of objects with fine texture (e.g., the least scratchy sandpaper). They fatigued these receptors by presenting to participants a high-frequency vibration. The idea here was that the high-frequency vibration would tire or cause adap- tation in the FAII receptors. After exposure to these vibrations, participants found the sur- faces of objects less “rough” than when they were not preadapted to these vibrations. A low-frequency vibration did not have these effects. Thus, this experiment supports the idea that our different mechanoreceptors contribute to different aspects of haptic perception (Pei, Hsiao, Craig, & Bensmaia, 2011). Reading Braille One skill in which haptic perception is vital is reading Braille. Braille is an alphabet system using easy-to-feel tactile letters instead of visual letters. It was developed by Louis Braille (1809–1852). Braille was a young boy in France when an accident robbed him of his sight. He struggled to use raised letters to read, which are diffi- cult to tell apart. Finding that this was the case for most blind individuals, Braille set out to develop a writing system that would be easier to read via touch. The sys- tem he developed nearly 200 years ago is still a criti- cal system for the blind today (see Figure 14.20 for an illustration of the Braille system). Interestingly, Braille was based on a night-writing system developed so that military orders could be given by the French army at night without having to alert enemies to their position. Braille is usually read using the index finger, which feels the dots raised above the surface of the page. Blind readers can now read Braille off the surface of a computer (Figure 14.21). It should be noted that in recent years, there has been a slight decline in the use of Braille, as computer technology allows text-to-voice translation, allowing blind individuals to hear instead of feel what they need to read. To read Braille, one must feel the dots raised above the page and perceive the number and pattern of the dots. As you can see in Figure 14.20, small differences in the number a b c d e f g h i j k l m n o p q r s t u v w x y z FIGURE 14.20 The Braille Alphabet These letters are raised off the page, so that blind readers can feel them. A fluent Braille reader can read almost half as fast as a sighted individual can read text. FIGURE 14.21 Person Reading Braille © iS to ck ph ot o. co m /p or to ka lis Exploratory procedures: hand movements made in order to identify an object 421 Chapter 14: Touch and Pain and pattern of dots indicate different letters. Thus, Braille is dependent on fine touch and the “what” system in the somatosensory cortex. Fluent visual readers may perceive many letters at the same time, but research shows that Braille readers feel fewer letters per unit of time than do visual readers seeing letters. On average, a competent Braille reader can read about 100 words per minute, somewhat less than the 250 words per minute sighted people can read (Mountcastle, 2005). As such, Braille reading tends to be slower than visual reading. In general, because our index fingers have the greatest number of mechanoreceptors, even compared with other fingers, Braille readers are able to read fastest by using their two index fingers together. Fluent Braille readers will also move the left hand down to the next line of text, even as the right hand continues to finish the line above (Lowenfield & Able, 1977). This research also suggests that natural left-handers can read Braille faster. Neuroimaging studies have revealed interesting reorganization of the brain in con- genitally blind individuals compared with individuals who became blind later in life (Voss & Zatorre, 2012). First, we see more white-matter fibers in the area of the somato- sensory cortex for congenitally blind individuals than for individuals who become blind later in life. Moreover, the occipital lobe reorganizes and receives input from the somatosensory cortex (and from the auditory cortex) in congenitally blind individuals. It is thought that this reorganization allows blind people greater area to devote to spatial processing deriving from nonvisual spatial modalities. Moreover, Braille readers show activity in the occipital lobe when reading (Voss & Zatorre, 2012). Although peo- ple who become blind later in life will show activity in the occipital lobe when touching objects, the earlier a person becomes blind, the more the occipital lobe is co-opted by other senses (Collignon et al., 2013). This research speaks to the relation of behavioral adaptation and neural reorgani- zation in the face of disruption of normal sensory systems. Thus, the brain does not waste the visual cortex when a person loses sight (due to damage to the eyes). Rather, the visual cortex is co-opted for spatial perception through other modalities. Reading by touch allows blind individuals the same opportunities in reading as does sight in normal individuals. It also gives blind people a distinct advantage: They can read in total darkness, without making a sound. Tactile Agnosia Tactile agnosia is a neurological condition caused by damage to the somatosensory areas of the parietal lobe. Tactile agnosia is defined as an inability to identify objects by touch, a seeming loss of haptic perception. For a patient to be diagnosed with tactile agnosia, he must show normal perception of texture, temperature, and pain and must show no deficits in motion. For example, a patient should be able to tell you that a surface is smooth and cold but not be able to identify it as an icepack. That is, the neu- rological problem must be one of identification rather than perception (Reed, Caselli, & Farah, 1996). An important characteristic of most patients with tactile agnosia is that only one hand is affected while the other hand is spared. This interesting observation is because the left hemisphere receives input from the right hand, and the right hemisphere receives input from the left hand. Thus, tactile agnosia is typically seen in only one hand. If the damage is to the left somatosensory cortex, tactile agnosia will be seen with the right hand, but not the left hand. If both hands show agnosia, a different neural cause is suspected (Gerstmann, 2001). Reed et al. (1996) studied a patient with tactile agnosia. E.C. was a 65-year-old woman who had tactile agnosia only in her right hand because of a left hemisphere Tactile agnosia: an inability to identify objects by touch 422 Sensation and Perception lesion caused by a stroke in that area. The patient had no difficulty identifying objects visually or with her left hand. However, when identifying objects by using only her right hand, she showed a massive deficit in object identification, though she could still identify textures, shapes, and characteristics such as smoothness. The patient also had cognitive knowledge of the shapes of objects. That is, she could describe the differences in textures among objects. However, when she felt objects with only her right hand (and could not see what they were), she was unable to identify most objects. The tactile agnosia, however, was not the only problem she had with her somatosensory system. She reported general numbness in her right hand, though thresholds and temperature sensitivity appeared normal. Thus, although her perception of basic sensory charac- teristics was not affected, her ability to identify objects was specifically impaired, thus classifying the diagnosis as agnosia. The Development of Haptic Perception The pattern of development of our sensations arising from the skin has not been studied extensively. There are studies that have determined the number of some of the different types of receptors in the skin over the life span. Some types of receptors reach their peak number before birth and actually start to decline in number before birth. Others reach their peak early after birth and decline throughout life. It should not be any surprise that various types of thresholds to skin sensations decline throughout the life span (Bleyenheuft & Thonnard, 2009). The more interesting studies of the development of touch have had to do with haptic perception. If you recall, studying infant visual perception has been difficult, and often we can only study visual perception starting at a few months. Streri, Lhote, and Dutilleul (2000) were able to study haptic perception in newborns. The average age of the infants in the study was 44.6 hours—that is less than 2 days old! Streri and colleagues used a haptic version of the habituation and dishabituation paradigm we discussed earlier (see Chapter 6 for an example). It appears that, because we use our hands to explore, our hands prefer novel objects just as our eyes do. An object was placed in the infant’s hand in a way that the infant could not see the object. Eventually, the infant let go and the researchers measured how long the infant held the object. On the first trial, the infant held the object for about 30 sec- onds. When the same object was placed in the hand, the infant held the object for a shorter time, under 10 seconds by the third trial. This finding could be due to habit- uation or fatigue. In the critical trial, a new object was put into the infant’s hand. The new object was carefully selected to differ only in the contours of the shape and no other general touch feature. The researchers hoped to show that haptic percep- tion was driving the results. With the new object, holding time increased to about 20 seconds, showing that habituation does explain the results. Given the careful choice in the object, a cylinder and a prism, these results also suggest that newborns have some haptic perception (Streri et al., 2000). Haptic perception continues to develop and children get better at using this sensory ability over the first few years of life (Bleyenheuft, & Thonnard, 2009). TEST YOUR KNOWLEDGE 1. Appraise haptic perception and how it differs from simple touch perception. 2. Define exploratory procedures and explain how they contribute to haptic perception. 423 Chapter 14: Touch and Pain THE VESTIBULAR SYSTEM: THE PERCEPTION OF BALANCE 14.6 Illustrate how the vestibular system operates to perceive head rotation and movement and interact with other senses to help us maintain balance. We are including the vestibular system in this chapter, although the vestibular system is not directly related to the somatosensory system, because this system deals with the movement of our bodies. The vestibular system may be most linked to the auditory sys- tem, with which it shares a number of anatomical features, but it interacts with several sensory systems, most particularly our proprioceptive system. Movement is essential to any animal, including humans. We are constantly in motion. As tempting as it is to remain in bed when your alarm clock goes off at 6 a.m., you must stand up and leave your bedroom. Once in the kitchen, you may need to bend down to pick up the coffee beans from the bottom shelf to brew yourself some coffee. After that, you may have to reach up to get a coffee cup from the top shelf. All of these ordinary daily activities require balance. Your weight shifts as you perform each of these tasks, and your ability to counter gravity changes as well. Balance is even more important in the sports we play. Consider a downhill skier who must keep herself upright on thin strips of carbon fiber while hurtling down a steep and slippery mountain at 85 mph (137 km/h). Balance here is the critical difference between winning a gold medal and spending the next 6 months recovering from serious injuries in the hospital. Nothing is worse for a skier than a compromised vestibular system! The vestibular system is the sensory system responsible for the perception of bal- ance and acceleration, particularly of the head, and it is housed in the semicircular canals and otolith organs, both located adjacent to the inner ear. The semicircular canals are three tubes located in the inner ear responsible for the signaling of head rota- tion. The semicircular canals are filled with a fluid called endolymph. The otolith organs are responsible for detecting acceleration of the head and identifying when the head is being held at a tilted angle. Hair cells within the semicircular canals signal information about rotations of the head, and hair cells within the otolith organs signal information about acceleration in linear directions. The anatomy of the vestibular system is depicted in Figure 14.22. Note the proximity of the vestibular system to the cochlea of the ear; in fact, they are connected and share fluid. In addition, the same nerve that carries auditory information to the brain carries the vestibular information. This is why a head cold can sometimes affect both our hearing and our sense of balance, as both of these areas may be clogged. The semicircular canals are tubes aligned in a perpendicular way to one another. Each semicircular canal is filled with a liquid called endolymph. At the bottom of each canal is a chamber called the ampulla, which contains a structure called the crista. Inside the crista are the hair cells, the receptors. Similarly, in the otolith organs is a structure called the macula, which houses the hair cells that respond to changes in head orientation. When our head moves in any particular direction, the endolymph lags behind because inertia and gravity cause a relative movement in the opposite direction. Think of holding up a glass. If you hold it straight, the water level is aligned with the glass. However, if you tilt the glass to the left, the water does not move with the glass but lags behind and tries to stay aligned with gravity. The water stays aligned with the floor but Vestibular system: the sensory system responsible for the perception of balance and acceleration, housed in the semicircular canals and otolith organs, both located adjacent to the inner ear Semicircular canals: three tubes located in the inner ear responsible for the signaling of head rotation Otolith organs: organs responsible for detecting acceleration of the head and identifying when the head is being held at a tilted angle Endolymph: fluid that fills the semicircular canals Ampulla: the structure at the base of each semicircular canal that contains the crista Crista: the structure in the ampulla of each semicircular canal that contains the receptors Macula: the structure in the otolith organs that contains the receptors 424 Sensation and Perception Saccule Semicircular canals Cupula displacement Direction of acceleration Endolymph flow Otoliths Vestibular nerve Facial nerve Cochlear nerve Utricle Cochlea Gelatinous material (a) (b) (c) Hair cells Neurons FIGURE 14.22 The Anatomy of the Vestibular System A diagram of the semicircular canals and the otolith organs. (a) A diagram of the main structures of the vestibular system showing its location relative to the cochlea of the auditory system. (b) The receptors of the semi-circular canals. The receptors are embedded in the cupula and motion of the cupula is transduced by the receptors. (c) A detail of the otolith organs showing the hair cell receptors. Motion of the gelatinous material is transduced by the receptors. not the glass. Similarly, if you move your head, the endolymph lags behind and tries to stay aligned with gravity. If you move your head left, the endolymph lags behind, creating a relative motion to the right against the hair cells, which do keep up with the head. This relative motion bends the cilia on the hair cells in a manner very similar to what we saw with hearing, triggering a neural response. When the endolymph shifts, the hair cells bend, transducing a neural signal, which informs the brain about head motion. Because the three semicircular canals are perpendicular to one another, each transmits information about head rotation along a different axis. That is, by comparing the movement in each canal, the brain can make inferences about rotations to the left or right (yaw), in front or behind (roll), and up or down (pitch). Thus, the movement of endolymph in each canal can provide information about relative head movement in great detail. Meanwhile, the otolith organs are responsible for detecting linear acceleration. When you start sprinting down the track, hair cells in the otolith organs bend, trans- ducing a neural signal, thus providing you with information about motion and acceler- ation. Similarly, there are hair cells set to detect accelerations up and down and left and right. Thus, when you stand up or slide to the right, your brain will have information about these accelerations. If you have ever spent extended time at sea—sometimes even a relatively short ferry ride will do it—you may have experienced the following illusion created by the vestibu- lar system. After you return to dry land after your 7-day cruise through the Caribbean, you may continue to experience a sensation of rocking or swaying while on what you know to be terra firma, such as in your own home. In the vast majority of people, these symptoms disappear after a few hours, though in some very rare cases, they may persist 425 Chapter 14: Touch and Pain for years. It is likely that these symptoms are caused by adaptation in the vestibular system. After a day or two at sea, you adapt to the constant motion caused by the rocking boat, but then when you return to land, it takes a while to readapt to land, leading to the temporary disorientation. The nerve fibers from the semicircular canals and the otolith come together in the vestibular nerve that then synapses in a brain stem area called the vestibular complex. The vestibular complex projects to several areas, including the parietal insular vestib- ular cortex, located in the parietal lobe. This area of the brain is thought to maintain a representation of head angle, critical for maintaining balance. Patients with damage to this area of the cortex may experience distortions of orientation, including vertigo, a sense of the world spinning when it is not, and illusory tilt; that is, they may feel that their heads are tilted when they are not. Indeed, some patients with damage in the pari- etal insular vestibular cortex may hold their heads at odd angles because it feels like the proper orientation to them. Thus, the vestibular system is an important sensory system in its own right and quite different from the more widely known and studied systems. Damage to it can have severe consequences, but we seldom attend to it unless there is something wrong, such as in the case of seasickness. This system combines with information from many other senses, including vision, proprioception, and even touch, to help us keep our balance in the world. TEST YOUR KNOWLEDGE 1. Explain and illustrate the function of the vestibular system. 2. Diagram the anatomy of the vestibular system, and explain how its anatomy is related to its function. EXPLORATION: Electroreception in Fish The bodies of all animals generate electricity. Think about the billions of neurons throughout our nervous systems sending electrical signals among one another. In the terres- trial environment, the electric fields generated by animals are incredibly weak, because air is such a poor conductor of electricity. This is why we need to put electrodes directly on the scalp to record electroencephalograms. No known land animal has a sensory system that picks up electrical output, all claims of extrasensory perception notwithstanding. However, as every worrying mother knows, water is much better at conducting electricity than air. (At the first sign of lightning, every parent considers it a sacred duty to get every child out of the pool at once.) Thus, the electrical output of the nervous systems of fish is detectable by other fish if they are in relatively close proximity and if those fish have evolved special electroreceptors. It turns out that many species of fish have sensory organs that allow them to detect the electric fields of other fish. Many sharks and rays, for example, are able to detect prey in complete darkness by homing in on their electric fields (Hughes, 1999). Think of Jaws closing in on you as you swim at night simply because you are deep in thought. Now you might not really want to go back to the beach. Electroreception in fish can be considered either passive or active. Passive electroreception means simply that an organ- ism can detect the electric fields generated by other animals. It is analogous to our hearing—we hear the sound gener- ated by objects in our environment. Active electroreception means that the fish generates its own electric field around itself and then senses disturbances to that electric field. This Electroreception: the ability to detect electric fields, seen in many species of fish Passive electroreception: the ability only to detect electric fields Active electroreception: the ability to generate electric fields and then detect disturbances or changes to those electric fields caused by external events Vestibular complex: a brain stem area that receives input from the vestibular nerve and sends the information to the forebrain Parietal insular vestibular cortex: an area in the parietal lobe that receives input from the vestibular nerve and is responsible for the perception of balance and orientation 426 Sensation and Perception is analogous to the biosonar systems of bats and dolphins, animals that listen for echoes from their own calls. Electroreception involves detecting electric fields, but a number of electric fish have also evolved the ability to generate strong electric fields for the purpose of hunting and predation avoid- ance (Hughes, 1999). But our attention here focuses on electro- reception. Electroreception has evolved multiple times among many species of fish. Thus, our discussion includes much gener- alization. Electroreception has also been discovered in at least one species of aquatic mammals, the platypus (Pettigrew, 1999). The electroreceptors are located inside the scales along the surface of most fish, though variations exist. Electroreceptors are also found in the lateral line—a long row of mechano- receptors that stretch along the backs of many fish, which allow them to coordinate their rapid swimming movements. But the electroreceptors detect electric fields while the mech- anoreceptors detect the movement of water. The electrore- ceptors are known as ampullae of Lorenzini, after the Italian anatomist who first identified them. The ampullae are used for passive electroreception. A second class of receptors, called the tuberous receptors, are used for active electro- reception in those fish with this ability. These receptors are shown in Figure 14.23. In Figure 14.24, we see the mouth and teeth of a tiger shark (Galeocerdo cuvier). The pits or dots along its snout are ampullae of Lorenzini. The ampullae of Lorenzini and the tuberous receptors are tuned to electric fields of different frequencies. In terms of electricity, frequency means variations in voltage across time. Batteries, for example, produce direct current (DC), in which voltage does not vary across time. However, the electricity in our houses and offices is alternating current (AC), in which the voltage varies at a specific rate or frequency. Biological organisms tend to produce AC. The ampullae of Lorenzini are tuned to DC and to very low frequency AC, whereas tuberous receptors are tuned to much higher AC frequencies (Fields, 2007; Hughes, 1999). As such, the two types of receptors are tuned to very different kinds of electrical events in the water. The ampullae of Lorenzini are tuned to detect the weak volt- age produced by the movements and body functions of other nearby fish. The tuberous receptors, however, are tuned by evolution to detect the fields generated by specific electric discharges by that fish. That is, just as bats listen for their own echoes, the tuberous receptors sense the changes in the AC produced by the fish itself. Fish equipped with tuberous receptors can then detect the differences between smaller fish, which might mean food, and bigger fish, which they want to avoid. In general, these electric fields are not that powerful and would give a human hand only a mild shock. Until quite recently (the 1950s), scientists did not know that fish had this electric sense and puzzled over the pur- pose of these obvious sense receptors. Because electric senses are not feasible in land animals such as ourselves, it was hard to imagine that animals such as tiger sharks could be so dependent on this completely alien sense. This is not to say that humans were unaware of the electric abilities of fish. Egyptian fishermen have feared the Nile catfish for thousands of years, and Amazonian peoples have feared the electric eel for presumably at least as long. But it really was not until the 20th century that we determined the relation between these sensory receptors and electricity. FIGURE 14.23 Electroreception (Hughes, 1999) The ampullae of Lorenzini and the tuberous receptors. FIGURE 14.24 Tiger Shark The mouth and teeth of a tiger shark (Galeocerdo cuvier). The pits or dots along its snout are ampullae of Lorenzini. We probably do not want to mess with this guy. © iS to ck ph ot o. co m /S te ve nB en ja m in Ampullae of Lorenzini: the organ that contains the hair cells that detect electric fields, used in passive electroreception Tuberous receptor: the organs that contain the hair cells that detect electric fields, used in active electroreception 427 Chapter 14: Touch and Pain APPLICATION: Phantom Limbs and Phantom Limb Pain Consistently, as teachers, we find that one of the phenom- ena our students are most interested in is phantom limb syndrome. This condition occurs when people who have lost a limb feel as if it is still there, either through pain or proprioception. In people who have lost limbs through warfare, accidents, or surgery, false proprioception can occur, in a condition known as phantom limb syndrome. Phantom pain occurs when a person feels pain in a part of the body that is missing. Phantom limb syndrome refers to the continued but illusory sensory reception in a missing appendage. Some amputees have the sense that they feel the position of a missing limb. That is, an amputee may feel as if his arm is resting at his side, in much the way we know our arm is resting at our side even if our eyes are closed. They can even control and move the missing limb as they would their remaining limb and feel it move (Walsh, Long, & Haggard, 2015). Phantom limb syndrome is likely the result of continued activity in the regions of the brain that would have responded to sensory input from the missing limb. Of course, phantom limb feelings can be quite dis- tressing at first for the patient. However, phantom limb proprioception, confusing as it must be at times, has a benefit. Phantom limb proprioception is useful in help- ing an amputee adjust to life with a prosthetic (Weeks, Anderson-Barnes, & Tsao, 2010). Patients with phantom limb proprioception quickly associate the feeling they have in the missing leg to the prosthetic. Amputees who lack phantom limb proprioception take longer to use the prosthetic (Weeks et al., 2010). Refer all the way back to Figure 14.13. The somatotopic map showed how the body is laid out across the somato- sensory cortex. Examining this map closely, you will see that the representation of the face is very close to the repre- sentation of the hands, and the representation of the face is not near the top of the neck. This organization turned out to be important in determining some of the mysteries asso- ciated with phantom limb syndrome. Patients with phan- tom limb syndrome may experience sensation in a missing hand when they are touched on the face. They may also experience sensation in a missing hand when touched on what remains of their arm. Ramachandran and Hirstein (1998) hypothesized that this is because the hand and face are next to each other on the somatosensory cortex, just as the area for the remaining limb is next to the hand in the somatotopic map. In their view, the area normally respon- sible for feeling on the face starts to draw on areas that are otherwise dormant because they are not receiving signals from a hand no longer present. However, because the rest of the brain is functioning on the assumption that the area responsible for the hand is perceiving sensation from the hand, the touching of the face causes the experience of having one’s missing hand touched. Initially, this sensa- tion may be disturbing to patients, but they get used to it. In some cases, however, patients may continue to have illusions of feeling in their missing hands. This explana- tion is very similar to the explanation for Aristotle’s illu- sion (review ISLE 14.2). You can also see Professor Ramachandran talking about phantom limb pain in ISLE 14.6. Unlike proprioception in which the experience of a phan- tom limb can help an amputee, real problems occur when people experience pain in missing limbs. Phantom limb pain refers to experiencing pain in the lost limb. Because there is no limb, amputees may be very confused when they experience pain in a missing appendage. Indeed, sometimes the pain can be excruciating, made all the worse by the amputee’s knowledge that the limb is not there. If the lost limb is one of the arms, 51% of patients experience phantom limb pain, whereas phantom limb pain is experienced by more than 70% of patients if a leg is lost (Kooijman, Dijkstra, Geertzen, Elzinga, & van der Schans, 2000). One way that phantom limb pain is experienced is as if the amputated limb is painfully cramped. Think of a never-ending muscle cramp in your calf that nothing you can do will relax. There is a general consensus that the pain being experienced is a top-down phenomenon that is generated by the brain trying to deal with the missing ISLE 14.6 Professor Ramachandran and Phantom Limb Syndrome Phantom limb syndrome: continued but illusory sensory reception in a missing appendage Phantom limb pain: refers to experiencing pain in a limb that has been amputated 428 Sensation and Perception input information. But there is also growing evidence that the pain, like the phantom limb syndrome itself, is generated by the remaining peripheral neurons (Vaso et al., 2014). Traditional drug therapies for pain have not proven terribly effective (Perry, Mercier, Pettifer, Cole, & Tsao, 2014), and researchers have searched many alter- native methods to help their patients. Given the difficulty in treating phantom limb pain, researchers have tried several alternative therapies to help the sufferers. Let us look at a couple of these therapies that relate to topics we discussed earlier. One interesting therapy requires a combination of using vision and proprioception to help ease the pain of phantom limb pain sufferers. Consider a case in which the patient has lost a hand. In this therapy, the patient will put both arms into a box and look at the side with the remaining limb. On the wall between the two limbs there is a mirror facing the remaining limb. The person cannot see the amputated limb but in the mirror the remaining limb is reflected to look to be in the position of the amputated counterpart. While in this mirror box, the patient moves his remaining limb around. It will appear as if the amputated limb is mov- ing. Even people with both limbs intact may experience the movement of the visible limb as the movement of the hidden limb (you can try this at home). The visual input seems to override proprioception in this situa- tion. That is, seeing a limb move is more important to our experience than the proprioceptive signals from limb and muscle. If you try this yourself, you will feel your hidden hand move, even though it is at rest and it is really the other hand moving. Patients can now feel as if the amputated hand is moving even though it is no longer present. So how does this strange phenomenon work as a therapy? The exact movements done will depend upon the type of pain being experienced, but let us use the example of a person experiencing a painful cramp in an amputated hand. The amputee will clench the remaining hand and then slowly open it. The amputee will see his missing hand open and may even feel the missing hand relax, reducing the pain (Chan et al., 2007; Foell, Bekrater- Bodmann, Diers, & Flor, 2014). Repeating this process may eventually eliminate the pain altogether. You can see a video of the use of mirror therapy for phantom limb pain and instructions on how to make your own mirror box in ISLE 14.7. Another approach that has recently emerged is to use virtual reality (refer back to the Application section in Chapter 7). The approach of this therapy has some sim- ilarities to the mirror therapy in that the missing limb is replaced visually, only in this case the limb is replaced digitally rather than through a mirror. Researchers use virtual reality to replace the amputated limb while the patient is in a virtual world. The patient can see and actu- ally use the missing limb to accomplish tasks in the virtual environment. Depending on the type of pain, the ampu- tee might perform different tasks designed to reduce that pain. Using the cramp example from earlier, the amputee might be asked to mentally relax the amputated limb and will see a virtual copy of the limb relax. As in the mirror box, she will feel as if her own missing limb is performing the actions of the virtual reality limb. Early results sug- gest that the use of visual feedback can help the patient experience reduced pain (Perry et al., 2014). With the new virtual reality goggles discussed in Chapter 7, it might be possible for patients to do their virtual reality exercises at home as well as in the clinic. It might also help bring down the cost of this technically demanding form of therapy. ISLE 14.7 also has video about a person using vir- tual reality to deal with phantom limb pain. There are several other emerging therapies for this serious issue, including directly injecting painkillers into the dorsal root ganglion (Vaso et al., 2014), but the therapies we’ve discussed are highlighted because of the interesting feature of how the senses interact. In both cases, it seems that vision dominated the information from proprioception. We are visual animals, and sometimes vision can override information from our other senses, a phenomenon called visual capture (Pavani, Spence, & Driver, 2000). Visual capture occurs when there is a conflict between the infor- mation from two senses and our experience follows the visual input rather than the other sense. In the therapies discussed, either through the mirror or virtual reality, visual input of a missing limb can overcome the proprio- ceptive information that the limb is absent. Visual capture does not always happen, and we need all of our senses, but in these cases, the use of visual capture has been exploited for therapeutic purposes. ISLE 14.7 Treatment of Phantom Limb Pain Visual capture: circumstances where visual input can dominate the input from other senses when they conflict 429 Chapter 14: Touch and Pain CHAPTER SUMMARY 14.1 Classify how touch perception is actually a num- ber of senses served by a number of receptor systems. Our somatosensory system starts with receptors embedded in the skin as well as in our muscles, joints, and tendons. Unlike audition and vision, touch requires direct physical contact with the object to be perceived. The somatosen- sory system is also composed of several distinct senses. Mechanoreceptors are embedded in our skin. The epider- mis is the outer layer of the skin, housing touch receptors, and the dermis is the inner layer of skin, also housing touch receptors. Mechanoreceptors are sensory receptors in the skin that transduce physical movement on the skin into neu- ral signals, which are sent to the brain. They tell us what we are touching and how to respond to it. Mechanoreceptors are divided into four types: SAI mechanoreceptors, SAII mechanoreceptors, FAI mechanoreceptors, and FAII mechanoreceptors. SA refers to slow-adapting, whereas FA refers to fast-adapting. “I” mechanoreceptors have small receptive fields, whereas “II” mechanoreceptors have larger receptive fields. Proprioception is the perception of the movements and position of our limbs. Proprioception gives us impor- tant information about the placement of our bodies. Proprioception is subserved by three kinds of receptors, muscle spindles, joint receptors, and Golgi tendon organs. Interestingly, an illusion of proprioception may occur in amputees. When this occurs, they are said to have phan- tom limb syndrome. Thermoreceptors are the sensory receptors in the skin that signal information about the temperature as measured on the skin. Cold fibers are thermoreceptors that fire in response to colder (30 °C and below) temperatures as mea- sured on the skin. Warm fibers are different thermorecep- tors that fire in response to warmer temperatures (above 36 °C) as measured on the skin. When the temperature is abnormally cold, such as when liquid nitrogen is placed on the skin, this extremely cold stimulus will also activate the warm fibers, leading to a perception of heat (and pain). Pain is the perception and the unpleasant experience of actual or threatened tissue damage. As unpleasant as it is, pain is a warning to avoid dangerous objects and protect oneself against greater damage. Nociceptors are sensory receptors in the epidermis and dermis that, when activated, cause us to feel pain. Nociceptors are divided into A-delta fibers, which conduct signals rapidly and respond to both heat and pressure, and C-fibers, which react more slowly and respond to pres- sure, extreme degrees of either heat or cold, and toxic chemicals. 14.2 Sketch the two main pathways from receptors in the skin to the somatosensory cortex in the brain. The dorsal column–medial lemniscal pathway is the pathway for the mechanoreceptors (tactile perception) and the proprioceptors (muscle position). It travels up the spinal column on the ipsilateral side and crosses to the contralateral side in the medulla. The other tract for the somatosensory system is the spinothalamic pathway, which conducts information from the nociceptors (pain) and the thermoreceptors (temperature). It travels up the contralateral side of the spinal column. The somatosen- sory cortex is an area in the parietal lobe of the cerebral cortex devoted to processing the information coming from the skin senses. It houses maps of the skin for the differ- ent skin senses. Thus, a touch to an area of the body will cause activation in these areas of the cortex. The maps are also scaled for the number of receptors. Thus, there is more space in the somatosensory cortex devoted to the fingers than to the upper back. Gate control theory is a model that allows for top-down control of the pain signal coming up the spinal cord. Indeed, activation in the anterior cingulate cortex of the frontal lobe may initiate a sequence that reduces pain via the gate control mechanism. 14.3 Examine the role of endogenous opioids in con- trolling pain perception. Pain may also be inhibited by endogenous opioids, which are chemicals produced by the body that reduce pain throughout the body. Sensation and Perception430 14.4 Contrast the difference between pain and itch and how they interact. Recent research suggests that itch perception is separate from pain perception. Itch induces scratching, whereas pain induces movement away from the source of pain. Pruriceptors are receptors in our skin that respond to mild irritants by producing itch sensations. 14.5 Assess the role movement plays in enhancing the perception of touch, allowing us to perceive objects. Haptic perception is the active use of touch to identify objects. Haptic perception includes integrating infor- mation from mechanoreceptors, proprioceptors, ther- moreceptors, and perhaps nociceptors as well. Haptic perception integrates movement and touch to determine the shapes and identities of objects, known as exploratory procedures. Braille reading uses tactile perception of raised dots rather than visual letters. It allows completely blind individuals to read. Braille readers have been shown to have more area in the occipital lobe devoted to spa- tial and somatosensory touch. Tactile agnosia may occur after brain damage to the somatosensory cortex. People with tactile agnosia can still feel texture and shape but cannot identify objects by touch. 14.6 Illustrate how the vestibular system operates to perceive head rotation and movement and interact with other senses to help us maintain balance. The vestibular system is the sensory system respon- sible for the perception of balance and acceleration. It is housed in the semicircular canals and otolith organs, both located adjacent to the inner ear. The semicircular canals are three tubes located in the inner ear respon- sible for the signaling of head rotation. Otolith organs are responsible for detecting acceleration of the head and identifying when the head is being held at a tilted angle. Both project to the parietal insular vestibular cortex. REVIEW QUESTIONS 1. What are mechanoreceptors? What are the four kinds of mechanoreceptors, and what kinds of stim- uli do they respond to? 2. What is proprioception? Where do we find the receptors for proprioception? 3. What is phantom limb syndrome? How does it manifest itself in both proprioception and the pain domain? 4. What is thermoreception? What two fibers project thermoreception information? What is the difference between them? 5. What are nociceptors? What is the difference between A-delta fibers and C-fibers? How do noci- ceptors differ from pruriceptors? 6. What is a homunculus? What does it tell us about the organization of the somatosensory cortex? What are the subareas of S1 (the primary somatosensory cortex)? 7. What is gate control theory? What anatomical regions are associated with it? How does it differ from regulating pain with opioids? 8. What are exploratory procedures? What do they tell us about haptic perception? 9. What is the vestibular system? What are the impor- tant anatomical correlates of the perception of bal- ance in the vestibular system? 10. What is electroreception? What species use it? What is the difference between passive and active electroreception? Chapter 14: Touch and Pain 431 PONDER FURTHER 1. The sensations that arise from the skin are unique in that sensations can arise passively (i.e., a stimulus comes in contact with the skin) or actively (i.e., we can reach out with our hands and interact with objects). How does this unique feature of our skin senses play a role in how these sensations are experienced? 2. Create a categorical system to illustrate the different sensations that arise from the skin. KEY TERMS Active electroreception, 425 A-delta fibers, 409 Afferent fibers, 407 Ampulla, 423 Ampullae of Lorenzini, 426 Analgesia, 417 Anterior cingulate cortex, 416 C-fibers, 409 Cold fibers, 408 Contralateral, 412 Crista, 423 Dermis, 403 Dorsal, 412 Dorsal column–medial lemniscal pathway, 412 Dorsal horn, 415 Dorsal root, 411 Dorsal root ganglion, 411 Electroreception, 425 Endogenous opioids, 417 Endolymph, 423 Epidermis, 403 Exploratory procedures, 420 FAI mechanoreceptors, 404 FAII mechanoreceptors, 404 Gate control theory, 415 Golgi tendon organs, 407 Haptic perception, 419 Homunculus, 414 Ipsilateral, 412 Joint receptors, 406 Macula, 423 Mechanoreceptors, 404 Meissner corpuscles, 405 Merkel cells, 405 Muscle spindles, 406 Nociceptive pain, 409 Nociceptors, 409 Otolith organs, 423 Pacinian corpuscles, 406 Pain, 409 Parietal insular vestibular cortex, 425 Passive electroreception, 425 Phantom limb pain, 427 Phantom limb syndrome, 427 Proprioception, 406 Pruriceptors, 418 Ruffini endings, 405 SAI mechanoreceptors, 404 SAII mechanoreceptors, 404 Semicircular canals, 423 Somatosensory cortex, 413 Somatotopic map, 414 Spinothalamic pathway, 413 Substantia gelatinosa, 415 Tactile agnosia, 421 Thermoreception, 408 Thermoreceptors, 408 Tuberous receptor, 426 Ventral, 412 Ventral posterior nucleus of the thalamus, 413 Ventral root, 411 Vestibular complex, 425 Vestibular system, 423 Visual capture, 428 Warm fibers, 408 Sensation and Perception432 Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review, study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning. Learning Objectives Digital Resources 14.1 Classify how touch perception is actually a number of senses served by a number of receptor systems. How Do You Amputate a Phantom Limb? V. S. Ramachandran’s Tales of the “Tell-Tale Brain” 14.2 Sketch the two main pathways from receptors in the skin to the somatosensory cortex in the brain. Feeling Better: Separate Pathways for Targeted Enhancement of Spatial and Temporal Touch Brain Implant Restores Sense of Touch to Paralyzed Man 14.3 Examine the role of endogenous opioids in controlling pain perception. How Is Pain Influenced by Cognition? Neuroimaging Weighs In In Pain Thou Shalt Bring Forth Children: The Peak-and-End Rule in Recall of Labor Pain Pain Really Is All in Your Head and Emotion Controls Intensity 14.4 Contrast the difference between pain and itch and how they interact. The Mystery and Power of the Itch Scientists Identify “Itchy” Neurons in Mice 14.5 Assess the role movement plays in enhancing the perception of touch, allowing us to perceive objects. 14.6 Illustrate how the vestibular system operates to perceive head rotation and movement and interact with other senses to help us maintain balance. 2-Minute Neuroscience: Vestibular System Platypus Hunts With “Sixth Sense” Six-Foot Electric Eel Kai Tirkkonen/The Image Bank/Getty Images 15Olfaction and Taste Will & Deni McIntyre / Science Source LEARNING OBJECTIVES 15.1 Examine the role of the olfactory system in detecting harmful and helpful substances. 15.2 Describe the anatomical and physiological bases for olfaction. 15.3 Appraise the nature of olfactory perception, olfactory imagery, and olfactory illusions. 15.4 Explain the anatomical and physiological bases for taste perception. INTRODUCTION Karen is a guidance counselor and a swim coach at a local high school. She was a competitive swimmer and still likes to get into the pool and demonstrate good form to her athletes. She’s happily married to a science teacher and track coach, and they have a 15-year-old daughter who excels at playing the flute. Karen and her family are certainly a successful family, happy and secure. However, Karen suffers from migraines. Migraines are terrible headaches that bring pain, nausea, and discomfort and can inca- pacitate a person for days. Migraines can be triggered by stress, changes in weather patterns, pollution, and, critical to our discussion, some odors. Karen seldom gets migraines anymore, because she has been able to remove most of the triggers from her daily life. However, when her husband and daughter took her out to dinner for her 50th birthday, a perfume being worn by a woman at a nearby table triggered a migraine, ruining her 50th birthday party. We can wonder why certain odors, particularly fra- grant perfumes, induce migraines in some people, whereas for others, the smell is pleas- ant enough to swathe one’s body in it. Perfume-induced migraines have become such a problem for some people that migraine sufferers have attempted to use the Americans with Disabilities Act to prevent coworkers from wearing strong, migraine-inducing per- fumes (Oltman, 2009). In contrast, think of the smell of one of your favorite foods or drinks—the aroma of coffee, the flavor of chocolate melting in your mouth, the smell of bananas wafting off a frying pan. In contrast to Karen, you might also find some perfumes to be very pleas- ing. Think of the perfume your girlfriend wears or the cologne your boyfriend wears. Even mentioning these odors and flavors in a textbook may induce some of you to take a break from reading and head to the kitchen for a snack. Indeed, odors may induce cravings for particular foods (Herz, 2007). Cravings may often seemingly force people to desperately search for an open pizzeria, regardless of the time of day. And for many of us, nothing is more pleasurable than eating our favorite foods. ISLE EXERCISES 15.1 Brain Area for Olfactory Bulbs 15.2 Posterior Piriform Cortex 436 Sensation and Perception In this chapter, we examine the sensory apparatus underlying our senses of olfaction (smell) and gustation (taste), how they interact, and what functions they serve. We also discuss the trigeminal nerve system, which is important to both smell and taste and also receives input from the somatosensory system. Together olfaction and gustation are considered chemical senses because their role is to detect chemicals in the environment. Olfaction brings airborne chemicals to our attention, whereas gustation alerts us to the chemical compositions of substances brought into our mouths. Similar to vision and audition, olfaction is a distal sense, detecting objects that may be some distance from a person. However, gustation is sim- ilar to the somatosensory systems, in that the taste must be in direct contact with the taste receptors in the mouth for gustation to detect it. Thus, when we encounter cof- fee, our olfaction system detects the chemicals emanating from the beverage before we ingest it, and our sense of taste samples it as it passes through our mouth. The chemical senses may be the oldest senses in an evolutionary framework. Many unicellular organ- isms have evolved to detect chemicals, either food or toxins, in their environment. OLFACTION 15.1 Examine the role of the olfactory system in detecting harmful and helpful substances. Olfaction refers to our ability to detect odors. Odors are the perceptual experiences that derive from the detection of odorants, which are airborne chemical molecules. Odorants are volatile chemicals; that is, in order to be smelled, an odorant must be able to float through the air and thus pass into our nose. Odorants must be repellent to water and relatively small molecules in order to be detected by our olfactory systems. In this way, our sense of smell is an early warning system, allowing us to detect potentially harmful or helpful substances before we come into direct contact with them. Our olfactory system does not respond to all airborne chemicals. Think of carbon monoxide, a toxic chemical produced by car engines. We cannot smell this odor, so we are unaware of the buildup of carbon monoxide. This is why sitting in a car in the garage with its engine running is dangerous. The engine produces carbon monoxide, which cannot escape because of the enclosed condition. Because we do not smell the carbon monoxide, it can build up to lethal levels without any warning. Natural gas is also odorless. But when it is used to heat homes, a chemical is added to give it a strong, unpleasant odor. Why would you want to make something smell bad? Well, a natural gas leak can be dangerous, but if the gas is odorless, it will not be detected. With the additive, a leak will smell like rotten eggs, something that surely gets a person’s atten- tion. Other unpleasant odors allow us to avoid dangerous situations. Rotting meat may contain toxic bacteria that can make us very sick. Thus, we evolved to find this smell putrid and unpleasant. In some cases, however, cultures may find ways to treat rotting meat to avoid sickness, in which case people learn to appreciate the smell rather than be repulsed by it (Herz, 2007). Many plants and animals emit molecules into the air, some of which are intended to be detected by themselves, other members of their own species, or other species (Figure 15.1). Many animals emit pheromones, which may indicate their mating status to other members of their species. In other cases, animals may use odorants to mark their terri- tory. Furthermore, in some cases, plants and animals may deliberately produce odorants for self-defense. Think of the characteristic odor of skunks. Skunks deliberately produce these odorants to deter predators, which are repelled by skunks’ odor (see Figure 15.1). Olfaction: the sense of smell Gustation: the sense of taste Odors: the perceptual experience of odorants, which are airborne chemical stimuli Odorants: molecules our olfactory system responds to when we detect them in the air 437 Chapter 15: Olfaction and Taste TEST YOUR KNOWLEDGE 1. What is olfaction, and how does it differ from gustation? 2. What is the difference between an odor and an odorant? How might animals use odors to communicate with other members of their species or other species? THE ANATOMY AND PHYSIOLOGY OF THE OLFACTORY SYSTEM 15.2 Describe the anatomical and physiological bases for olfaction. The Nose We have two nostrils in our nose, which serve as the entranceway into the nasal cavi- ties. The nostrils are separated by a wall of cartilage called the nasal septum. Damage to the nose, such as from a punch or a hard fall, can cause a deviated septum, in which the wall of cartilage is no longer straight. A deviated septum can interfere with © iS to ck ph ot o. co m /L yn n_ By st ro m © iS to ck ph ot o. co m /L au riP at te rs on © iS to ck ph ot o. co m /s tr ay _c at © iS to ck ph ot o. co m /c ob ra ph ot o © iStockphoto.com /atoss © iStockphoto.com /sam cam FIGURE 15.1 Odors Examine these photographs. Each is associated with a characteristic odor. Consider each object and think about the odor of that object. Is it pleasant or unpleasant? Is it weak or strong? Nasal septum: the wall of cartilage that separates the nostrils 438 Sensation and Perception both proper breathing and the sense of smell. The nasal septum may also be punctured or perforated. For example, chronic cocaine use can cause holes to form in the septum, which can also interfere with breathing and olfac- tion. In human beings, the two nasal cavities are extremely close to each other, so they are essentially sampling the same air. Thus, there is no real analogy to the binocular vision our two eyes can achieve or the ability of our audi- tory system to integrate across ears, such as in sound localization. Inside the nasal cavity, turbinates serve to disperse air. Turbinates are bony knots of tis- sue that ensure that some air will be passed upward through a space called the olfactory cleft and land on an area of tissue called the olfactory epithelium. Olfactory recep­ tor neurons are located inside the olfactory epithelium, and these neurons serve as the transducers of the olfactory system. The olfac- tory receptor neurons are pretty far into the nasal cavity; indeed, they are just a couple of centimeters behind each eye. The air passes through the cleft and then rejoins the air being sent through the pharynx toward the lungs. Odorants from food find their way to the olfactory epithelium through a passage in the oral cavity in the back of the mouth. This is illustrated in Figure 15.2. The olfactory epithelium serves as the organ of transduction, taking chemical stimulation and transforming it into a neural signal. There is an olfactory epithelium at the top of each nostril, measuring about 1 to 2 cm2. In addition to the olfactory receptor neurons, the olfactory epithelium contains supporting cells and basal cells. Supporting cells provide metabolic supplies to the olfactory receptor neurons. Basal cells create olfactory receptor neurons. Interestingly, olfactory receptor neurons die out after about a month, so basal cells are continually resupplying the olfactory epithelium with olfactory receptor neurons. It is estimated that olfactory receptor neurons are completely regenerated every 28 days. The anatomy of the olfactory epithelium is shown in Figure 15.3. The olfactory receptor neurons have cilia extending into the mucus covering of the olfactory epithelium. These cilia contain the transducing elements of the cell on their tips. As a particular molecule of a particular odorant comes into contact with the tip of the cilium of the olfactory receptor neuron, a chain of chemical events is initiated, ending with an action potential leaving the olfactory receptor neuron along an axon heading toward the olfactory bulb. That is, the chemical triggers the cilia, causing a neural signal to begin. As few as seven molecules of the same odorant can trigger an action potential in an olfactory receptor neuron. There are approximately 20 million olfactory receptor neurons in the human nose. Although this is more sensory neurons than for any other human sensory modality, it is much fewer olfactory receptor neurons than seen in other species. Dogs such Olfactory bulb Olfactory cleft Turbinates Oral cavity Tongue Pharynx Cribriform plate Olfactory epithelium Nasal cavity Odorant molecules Odorant molecules in olfactory mucosa Olfactory bulb Basal cell Olfactory sensory neurons Olfactory receptors Glomerulus Mitral cell Tufted cell Olfactory epithelium Cribriform plate Olfactory mucosa Supporting cells Olfactory cilia Bowman’s gland FIGURE 15.2 Gross Anatomy (No Pun Intended) of the Human Nose Odorants enter the nasal cavity and make their way up to the olfactory epithelium. The turbinates, which are bones covered with epithelial tissue, keep the air circulating up toward the olfactory epithelium. Turbinates: bony knots of tissue that serve to disperse air within the nasal cavity Olfactory cleft: the channel at the back of the nasal cavity that funnels air up toward the olfactory epithelium Olfactory epithelium: a mucous membrane inside each nostril of the nose that contains the receptor cells for the olfactory system Olfactory receptor neurons: receptor cells located in the olfactory epithelium that detect specific chemicals in the air and transduce them into a neural signal 439 Chapter 15: Olfaction and Taste as bloodhounds and basset hounds have as many as 10 times more olfactory receptor neurons than we do. Grizzly bears may even have more olfactory receptor neurons than dogs. Pigs also have large concentrations of olfactory receptor neurons. The density of the olfac- tory receptor neurons in these species gives these ani- mals an enormously sensitive sense of smell. Grizzly bears can detect the aroma of meat from nearly 20 miles away (Figure 15.4). Bloodhounds can detect the pres- ence of a specific animal even if that animal passed by a location several days earlier. Pigs are trained to detect truffles buried underground. Species, such as pigs, bears, and dogs, that depend heavily on smell are called macro­ smatic. Humans, who are more dependent on vision and audition, are considered microsmatic. You can see some illustrations of different animals’ brains and the space devoted to olfaction on ISLE 15.1. It is estimated that humans have about 350 different kinds of olfactory receptor neurons. Each kind of olfactory receptor neuron responds to a relatively small class of odorants. When we compare the olfactory system with the visual system, we find that the 350 types of olfactory receptor neurons roughly correspond to the three cones and one rod in the visual system. Thus, the function of identifying smells is a very different process than how we identify color. Macrosmatic species may have as many as 1,000 different types of olfactory receptor neurons. Genes and Olfaction In 2004, neuroscientists Linda Buck and Richard Axel received a Nobel Prize for their groundbreaking work on the genetics of transduction of olfactory receptor neurons. FIGURE 15.4 Grizzly Bear Grizzly bears have a tremendously sensitive sense of smell. © iStockphoto.com /Zw illing330 ISLE 15.1 Brain Area for Olfactory Bulbs Supporting cells: cells that provide metabolic supplies to the olfactory receptor neurons Basal cells: cells that create olfactory receptor neurons Macrosmatic: species that are heavily dependent on their olfactory system FIGURE 15.3 Anatomy of the Olfactory Epithelium (a) The olfactory receptor neurons are located beneath the mucous layer of the epithelium. The olfactory epithelium houses three types of cells: olfactory receptor neurons, basal cells, and supporting cells. The olfactory receptor neurons are found beneath the olfactory epithelium. (b) The location of the olfactory bulbs on the bottom of the sheep brain. They are much larger in sheep than human brains but lie in very similar locations in the brain. Use your anaglyph glasses to view this image. Olfactory bulb Basal cell Olfactory receptor neurons Glomerulus Mitral cell Tufted cell Olfactory epithelium Cribriform plate Olfactory mucosa Supporting cells Olfactory cilia Bowman’s gland (a) (b) 440 Sensation and Perception In 1991, Buck and Axel published an important paper that described their discovery of a collection of genes that regulate the expression of different olfactory receptor neurons (Buck & Axel, 1991). They described a family of about 1,000 genes across mammalian spe- cies that are involved in genetic coding for olfactory transduction. However, in human beings, the majority of these genes are inactive. Indeed, it is estimated that only about 350 of these genes actually code for olfactory reception in humans (Malnic, Godfrey, & Buck, 2004). In other species, such as the macrosmatic species (e.g., dogs, bears) just described, far fewer of these genes will be inactive. In humans, the number of active genes is a predictor of individual differences in olfaction. Research suggests that when a person has more copies of a particular gene active, she may be more sensitive to the odorant that the gene maps onto. Moreover, if a gene that allows people to detect a pleasant odor is absent, foods with that odor may not be as appealing to them as they are to people in whom the gene is expressed (Menashe, Man, Lancet, & Gilad, 2003). Thus, for example, if you lack a gene that codes for lavender, you may not like certain perfumes. The Trigeminal Nerve Many odorants also have a second sensory component to them, which often may be described as a feeling. Some odors burn, whereas oth- ers soothe. Indeed, odorants such as ammonia (burning) and menthol (cooling) also cause reactions in the somatosensory system. This aspect of olfaction is mediated by the trigeminal nerve, which transmits information about the feel of an odorant (Figure 15.5). Trigeminal stimulation accounts for a number of the experiences associated with eating food as well. For example, the burning sensation of chili peppers is the result of trigeminal stimulation. It is also trigeminal stimulation that induces tears when we eat or cut fresh onions. TEST YOUR KNOWLEDGE 1. Describe the gross anatomy of the nose. 2. What is the trigeminal nerve? How does it bridge the gap between olfaction and the somatosensory system? The Pathway to the Brain The olfactory receptor neurons project axons through little holes in the base of the skull called the cribriform plate, a bone of the skull that separates the nose from the brain. Because of its location and its perforated surface, the cribriform plate is sus- ceptible to injury when the head receives a hard blow. Even a human punch can cause the cribriform plate to become fractured. When it is fractured, it may sever the axons coming from the olfactory receptor neurons, causing impairment to the sense of smell. In some cases, anosmia may develop. Anosmia (or smell blindness) is the inability to smell, usually caused by cribriform plate damage. Despite the regeneration of olfactory receptor neurons, a fractured cribriform plate may scar over, preventing the new neu- rons from projecting through it to the brain. Figure 15.6 shows the passage of axons from the olfactory receptor neurons to the first areas of olfactory processing in the brain. Sinus infection may also cause anosmia, either temporarily or permanently. We return later, in the Exploration section of this chapter, to the causes and consequences of anosmia. Microsmatic: species that are less dependent on their olfactory system Trigeminal nerve: a nerve that is associated with the feel of odorants; also known as the fifth cranial nerve Cribriform plate: a perforated section of skull bone that separates the nose from the brain; axons from olfactory receptor neurons pass through to allow olfactory information to enter regions in the brain Anosmia (smell blindness): the inability to smell, usually caused by cribriform plate damage Olfactory nerve (first cranial nerve): the axons of the olfactory receptor neurons that leave the nose and enter the olfactory bulb Olfactory bulb: a part of the brain just behind the nose; it is the first place in the brain where olfactory information is processed Opthalmic nerve Mandibular nerve Trigeminal nerve (cranial nerve V) Ethmoid nerve (nose) Ciliary nerves (cornea) Maxillary nerve Lingual nerve (tongue) Inferior alveolar nerve (teeth) FIGURE 15.5 Anatomy of the Trigeminal Nerve The trigeminal nerve carries information from somatosensory receptors in the nose to the thalamus. The experience of “heat” when eating chili peppers is caused by activation of the trigeminal nerve. 441 Chapter 15: Olfaction and Taste The axons of the olfactory receptor neurons converge to form the olfactory nerve, which exits through the cribriform plate. The olfactory nerve (also known as the first cranial nerve) consists of the axons of the olfactory receptor neurons that leave the nose and enter the olfac- tory bulb. The olfactory bulb is a part of the brain just behind the nose. It is the first place in the brain where olfactory information is processed. Once inside the olfac- tory bulb, the axons from the olfactory receptor neurons enter and synapse with dendrites in spherical structures called glomeruli. There are two types of dendrites in the glomeruli. One of the dendrites is from mitral cells, whereas the other is from tufted cells. Mitral and tufted cells have similar projections, but differ in that they respond to different odorants, and mitral cells are more likely to inhibit the olfactory process than are tufted cells (Nagayama, Takahashi, Yoshihara, & Mori, 2004). These two types of cells form the olfactory tract, which projects olfactory information from the olfactory bulb to other regions of the brain. The olfactory bulb is one of the most forward (ante- rior) parts of the human brain. It is found in humans just above the nose and just behind the eyes. It is also a primitive part of the brain, in the sense that our olfac- tory bulbs do not differ greatly from other mammalian olfactory bulbs. The beginning of olfactory processing occurs in the glomeruli. The cells in the glomeruli form an odorant map, organizing similarly structured odor- ants together. This means that odorants from chemicals with similar structures are processed adjacent to one another. This map is analogous to frequency coding in the auditory cortex or spatial mapping in the visual brain. It is worth noting that odorants with similar chemical structures may not have similar subjective odors. Nonetheless, processing in the olfactory bulb is organized by chemical structure. Information passes from the glomeruli to the mitral and tufted cells, which then project to structures beyond the olfactory bulb. The axons of the mitral cells and the tufted cells project further into the brain and synapse in a variety of locations in the brain. Chief among these projections are con- nections to the amygdala (an emotion area), the entorhinal cortex of the temporal lobe (a memory area), and the piriform cortex, also in the temporal lobe (Figure 15.7). Of particular interest is the piriform cortex, which is often considered the primary olfac- tory cortex. The piriform cortex is devoted exclusively to olfaction, but the entorhinal cortex is an important memory area of the brain, and the amygdala is an important emotion area of the brain. The fact that the olfactory bulb directly projects to these areas is critical in two aspects of olfaction: its association with memory and its association with emotion. The direct connection between the olfactory bulb and the entorhinal cortex is likely why a particular odor can elicit an involuntary autobiographical memory so quickly. The entorhinal cortex is deeply connected to the hippocampus, also a known memory area of the brain. Many people will report that a particular smell will evoke strong memo- ries, often of childhood, with striking regularity. For example, the smell of naphthalene (or mothballs) always reminds one of your authors of his grandmother’s apartment from when he was a child. Glomerulus Olfactory nerve (cranial nerve I) Mitral cell Olfactory sensory neuron Receptor Odorant molecules Cribriform plate Olfactory cilia FIGURE 15.6 Olfactory Pathway The pathway of odors from the olfactory receptor neurons to the brain. Glomeruli: spherical structures within the olfactory bulb where the olfactory tract forms synapses with mitral cells and tufted cells Mitral cells: neurons that start in the glomeruli of the olfactory bulb and project to other areas of the brain; respond to different odorants than do tufted cells. Tufted cells: neurons that start in the glomeruli of the olfactory bulb and project to other areas of the brain; they respond to different odorants than do mitral cells Olfactory tract: the pathway leading from the olfactory bulb to other regions of the brain Amygdala: an area of the brain in the limbic system, associated with the experience of emotion, particularly fear 442 Sensation and Perception Moreover, the direct connection between the olfactory bulb and the amygdala is most likely pivotal in the rapid negative associations that occur when we smell an odor in a negative situation. Even in microsmatic humans, one up-close interaction with an angry skunk will leave you dis- liking the musk of skunk for life. Similarly, one very positive experience with a particular perfume or cologne, and you may find that odor pleasant for life as well. The amygdala projects directly to the hypothalamus, which is critical in the regulation of activities such as hunger, thirst, and sexual desire. Representation Within the Piriform Cortex The piriform cortex is found in the temporal lobe. It is adja- cent to areas of the brain known as the limbic system, which are critical in such functions as emotion and memory. The piriform cortex has two main anatomical subdivisions: the anterior piriform cortex and the posterior piriform cortex (Kadohisa & Wilson, 2006). As the names suggest, the ante- rior piriform cortex is located in the front portion of the piriform cortex, and the posterior piriform cortex is located in the back portion of the piriform cortex. The anterior piriform cortex is associated with representing the chemical structures of odorants. That is, like the olfactory bulb, the anterior piriform cortex creates a map of odorants organized by their chemical structure. Moreover, neurons within the anterior region are narrowly tuned, responding to a very small range of odorant molecules and not others. In contrast, the posterior piriform cortex is associated with an odor’s quality, regardless of its chemical compo- sition. That is, the posterior piriform cortex groups together odors that smell similar to us, regardless of chemical origin. You can see the location of the piriform cortex in Figure 15.7. You can see an interesting simulation of the posterior piriform cortex on ISLE 15.2. The posterior piriform cortex represents the subjective qualities of odors rather than their chemical compositions. In the posterior piriform cortex, neurons are grouped by subjective similarities between odors. For example, odors that smell “smoky” are grouped together, regardless of the types of molecules that elicit them. Moreover, chem- icals that are similar in their molecular structures may be represented in different areas of the posterior piriform cortex if they elicit different olfactory experiences. In this way, the posterior piriform cortex serves a similar function to the extrastriate cortex in vision, processing the identities of odorants in the environment, just as areas such as V2 and V4 process the identity of visual stimuli. An interesting study by Howard, Plailly, Grueschow, Haynes, and Gottfried (2009) illustrates the difference between the anterior piriform cortex and the posterior piriform cortex. Howard et al. recorded brain activity via functional magnetic resonance imag- ing while participants were sampling three different odor categories (minty, woody, and citrus). The odors within each category (e.g., minty) were subjectively similar but had different chemical structures. The pattern of brain activity showed that the posterior piriform cortex was similarly activated for odors that were subjectively similar; that is, they all smelled minty. However, when looking at other regions of the brain responding to odors, such as the anterior piriform cortex, this pattern was not seen. Therefore, Howard et al. concluded that the posterior piriform cortex must be involved in identi- fying the qualities of odorants. Thalamus Hippocampus Hypothalamus Entorhinal cortex Amygdala Piriform cortex Olfactory tract Orbitofrontal cortex Olfactory bulb Olfactory receptor neurons in olfactory epithelium FIGURE 15.7 Areas of the Brain Receiving Input From the Olfactory Bulb The olfactory bulb is located just behind the nose. It sends information to the piriform cortex (located in the temporal lobe), the entorhinal cortex, and the amygdala. The piriform cortex projects to the orbitofrontal cortex. Entorhinal cortex: an area in the medial temporal lobe, associated with a number of memory functions Piriform cortex: an area in the anterior region of the temporal lobe that receives input from the olfactory bulb and is involved in olfactory processing; often considered the primary olfactory cortex Anterior piriform cortex: a structure located in the front portion of the piriform cortex that is associated with representing the chemical structures of odorants Posterior piriform cortex: a structure located in the back portion of the piriform cortex that is associated with an odor’s quality, regardless of its chemical composition ISLE 15.2 Posterior Piriform Cortex 443 Chapter 15: Olfaction and Taste The Orbitofrontal Cortex Another important region of the brain that has an olfactory function is the orbi­ tofrontal cortex, particularly the right orbitofrontal cortex. This area receives pro- jections from both the piriform cortex and the limbic system. It is thought that this area of the brain is critical in establishing the emotional nature of odors. Odors are also fundamentally an emotional and affective experience (Herz, Eliassen, Beland, & Souza, 2004). Certain odors (e.g., coffee, chocolate, certain perfumes) may elicit very positive emotions, whereas others (e.g., skunk, rotting garbage, excrement) may elicit very negative emotions. Thus, the affective values of odors are important, as positive odors elicit approach and negative odors afford distancing. Much of this cognition appears to be occurring in the orbitofrontal cortex. The orbitofrontal cortex is located just behind and above our eyes, in the very most frontal part of the frontal cortex. This area of the cortex appears to be critical in integrating olfaction and taste per- ception of foods, allowing us to enjoy some foods and reject others. The right orbito- frontal cortex is critical to the actual experience of an odor, that is, the “feeling” we get when we smell an odor, or what philosophers call its qualia (Li, Lopez, Osher, & Howard, 2010). TEST YOUR KNOWLEDGE 1. Describe the neural pathways from the olfactory nerve through higher order areas of the brain. 2. What are the functions of the anterior piriform cortex and the posterior piriform cortex? OLFACTORY PERCEPTION 15.3 Appraise the nature of olfactory perception, olfactory imagery, and olfactory illusions. Detection We reviewed the anatomy of olfaction in the last section. Now we cover olfactory psychophysics. Here, the questions concern exactly what our sense of smell can do. In detection, we want to know how much of an odorant must be in the air for people to detect that odorant. It turns out that the answer to this question depends on what the molecule is. Some odors can be detected in much smaller quantities than others. Of course, as concentrations of a particular odorant increase, so will the likelihood of detection. Thus, just as the amount of light necessary to see depends on which wave- length of light is available, different odorants require different concentrations for olfac- tory detection. So, although our detection varies greatly from odor to odor, we can still make some generalizations about odor detection. The amount of odorant in the environment is typically measured as a function of the number of molecules of that odorant per 1 million molecules of air. This is known as parts per million, or ppm for short. If you recall from Chapter 2, thresholds will vary depending on a number of characteristics. However, averaging over these characteristics, we can discuss the mean ppm required to detect a particular odor. To refresh the concept, an absolute threshold is the minimum amount of stimulus intensity required to elicit a perception of that stimulus. Different odorants require different concentrations to be detected. For example, we are extremely sensitive to the Orbitofrontal cortex: a part of the prefrontal cortex that appears to be critical in the emotional experience of odors and integrating olfaction and taste perception, among other functions 444 Sensation and Perception sweet smell of vanilla; it requires only 0.000035 ppm to detect it (Figure 15.8). However, the smell of acetone, which many of us know as nail polish remover, requires 15 ppm to detect (Devos, Patte, Rouault, Laffort, & Van Gemert, 1990). We can also look at the just-noticeable difference (JND) for any particular odor- ant, that is, how much more odor we must add to the air for us to detect a stronger concentration of that odor. This will also vary among odors. Identifying Odors Imagine the following scenario. You arrive at a friend’s house and smell a beautiful aroma of something cooking in the kitchen. It smells wonderful, and you are sure you know the aroma, but you cannot name it. What is that smell? It is familiar to you, but you cannot put a name to it. This altogether common phe- nomenon has been called the tip­of­the­nose phenomenon, very similar to the more common tip-of-the tongue phenomenon (Jönsson & Olsson, 2012; Jönsson & Stevenson, 2014), and it occurs when a person is familiar with an odor but cannot recall its name, despite feeling as if he can. If your host tells you that she is cooking a curry, you instantly recognize the aroma as being that of curry and wonder why you did not identify the aroma yourself in the first place (Figure 15.9). You can even try this out with friends. Open up spice bottles for them and see if they can name the spices by sniffing. You will be amazed at how few smells they will be able to name. But if you give them multiple choices, they will be very accurate (Jönsson & Olsson, 2012; Jönsson & Stevenson, 2014). As we have discussed, there is an intimate connection between olfaction and emotion. Odors elicit emotions, both positive and negative. There is also a close connection between olfaction and memory—some odors elicit specific memories. However, there seems to be a disconnect between olfaction and language, ren- dering it difficult to identify the names of certain odors, espe- cially when these odorants are encountered in an unusual context (Cain, 1979; Herz, 2007). Research shows that when people are given odors to sniff in a lab, they are poor at naming these odors (Figure 15.10). Indeed, in most experiments, most participants sel- dom approach naming even half of the odors (Jönsson & Olsson, 2012). However, this does not mean the odors are not familiar and known to the participants—the problem is in labeling them. This point is easy to demonstrate by giving participants a recognition test in which they must identify the correct name of an odor from among a list of alternatives. Under these circumstances, odor identification increases tremendously. In one study, de Wijk and Cain (1994) found that young adults’ odor naming was just about 40% correct, but performance increased to above 80% cor- rect when the participants were permitted to choose from a list of possible names. Discrimination in psychophysics means distinguishing between two stimuli. In most psychophysics experiments, the two stimuli to be discriminated will be closely FIGURE 15.8 Vanilla Ice Cream The extract known as vanillin that creates the vanilla smell and taste is an odorant we are very sensitive to. © iS to ck ph ot o. co m /8 93 50 95 FIGURE 15.9 Foods and Their Smells Would you recognize the aromas of these dishes? © iS to ck ph ot o. co m /m uk es h- ku m ar Tip­of­the­nose phenomenon: a phenomenon that occurs when a person is familiar with an odor but cannot recall its name, despite feeling as if he or she can 445 Chapter 15: Olfaction and Taste related in frequency or some other variable. We are good at discriminating odors even when we cannot name them. It is estimated that we can distinguish more than a thousand different aromas and that pro- fessionals (e.g., wine tasters and perfumists) can dis- criminate as many as 100,000 different odors (Herz, 2007). There is some debate as to the advantages experts have in the sense domain, but the ability to discriminate is clearly demonstrable. Research shows that experts are better able to identify odors by name and identify subcomponents within an odor. However, whether wine experts can reliably predict which wines are of higher quality has been an issue of some debate, often showing results quite embarrassing to those experts. Odor Imagery A fascinating question in olfaction is why so few people are able to experience olfac- tory imagery. By olfactory imagery, we mean the ability to experience the “smell” of a particular odor when that odor is not physically present. Imagery is different from hallucination because the person generates the imagery herself and knows that the image is present only internally. In the visual domain and the auditory domain, it is easy to experience a sensory experience known to be not actually present. Just sitting there reading this text, you can make a mental visual image what a Sicilian pizza looks like, and without so much as moving your lips or vibrating your vocal chords, you can imagine your favorite song. But try coming up with an olfactory image of the smell of that pizza or what your significant other’s perfume smells like. Most people report being unable to experience olfactory images (Djordjevic et al., 2008). Thus, unlike the senses of vision and audition, olfactory imagery is difficult for most of us. However, there is some research that supports the idea that some people are able to make olfac- tory images. One study showed that people who reported olfactory imagery showed brain activity in the piriform cortex when engaged in olfactory imagery. Control partic- ipants asked to make olfactory images did not show this activity (Djordjevic, Zatorre, Petrides, Boyle, & Jones-Gotman, 2005). Olfactory Illusions Our chapters on vision and audition abound with the concept of illusions. Illusions are situations in which what we perceive is not what is physically present. For example, after staring at a waterfall, we see the world moving in the opposite direction, even though we know this is not so. We hear scales continually becoming higher pitched even though the frequencies do not get higher. Olfactory illusions exist, though unlike visual and auditory illusions, we are seldom conscious of the dissociation of stimulus and perception (Stevenson, 2011). Because we are seldom aware of these olfactory illusions, they get less attention than visual and auditory illusions. Nonetheless, we consider a few olfactory illusions here. There are a number of illusions that arise from context effects. That is, the olfactory environment that surrounds a particular odorant changes the way that odorant is per- ceived. Lawless (1991), for example, used the odorant chemical called dihydromyrcenol. FIGURE 15.10 Identifying Odors in a Lab Phanie/A lam y Stock Photo 446 Sensation and Perception When this chemical is surrounded by woody-smelling odors, it is perceived as smelling like citrus. However, when it is surrounded by citrus smells, it is perceived as being woody in character. This is like some of the center-surround illusions in vision, in which the border of an image influences the perception of lightness or color of the image inside that border. In Lawless’s illusion, the surrounding odor influences the perception of a target odor. In a number of other cases, surrounding odors can affect the perception of a target odor (Stevenson, 2011). Verbal labeling can also cause olfactory illusions. In particular, the label for a per- ceived odor will often affect whether the odor elicits a positive or negative emotion. Imagine being presented with an odor and being told it was “aged parmesan cheese.” It is likely that your reaction might be positive. But if you were told the same odor was “vomit,” you might have a very different reaction. Indeed, in this case, the two substances have odorants in common, so it is likely in this case that you would be influenced by the verbal label. This was put to the empirical test by Herz and von Clef (2001), who presented odors with both a positive label and a negative label but examined the influence of which label was presented first. In this case, participants read both positive and negative labels, but Herz and von Clef reasoned that the first label would have a bigger impact on emotional reaction to the odor. For example, the odor of pine trees was presented with either the label “Christmas tree” first or the label “toilet cleaner” first. The independent variable was which label occurred first. Herz and von Clef found an outsized role for the first label. If the “Christmas tree” label was given first, the odor was rated more positively than if the “toilet cleaner” label was given first. Thus, the same odor with the same labels received a different pleasantness judgment depending on the order of the labeling. For this reason, Stevenson (2011) classified this effect as an olfactory illusion. Other sensory modalities can also influence the perception of odor. For example, Engen (1972) showed participants colored or noncolored liquids. Participants were more likely to report smelling something in the colored liquids even when there were no odor- ants present in the liquids. In effect, the visual perception of a colored liquid induced an olfactory illusion of an odor even though the liquid had no odorants in it. The final olfactory illusion we consider is olfactory rivalry (Stevenson & Mahmut, 2013). Olfactory rivalry is analogous to binocular rivalry, the illusion created by pre- senting one image to the left eye and a very different image to the right eye. In vision, we see only one of the images at any one time, but the perception of which image it is may vary, seemingly randomly, across time. In olfactory rivalry, one odorant is presented to one nostril, and a different odor is presented to the other nostril. For example, the left nostril might receive the smell of roses (phenylethyl alcohol), whereas the right nostril might receive the smell of “permanent markers” (butanol). Using this procedure, Zhou and Chen (2009) found an analog to binocular rivalry. There was a seeming randomized likelihood that people would smell the roses or the markers. People reported smelling one and then the other but not both at the same time. As with binocular rivalry, the perception of odor would switch over time from one to the other, also in a seemingly random fash- ion. Stevenson and Mahmut showed that this illusion can occur even when both nostrils are receiving both odors. TEST YOUR KNOWLEDGE 1. What is the tip-of-the-nose phenomenon? What does it tell us about the relation of olfaction to other functions in the brain? 2. What are olfactory illusions? Can you describe one and how perception is tricked? 447 Chapter 15: Olfaction and Taste TASTE PERCEPTION 15.4 Explain the anatomical and physiological bases for taste perception. M. F. K. Fisher wrote a number of books about cook- ing and eating throughout her illustrious career as a writer. Born in Michigan, Fisher lived in France and California and produced some of the most delectable writing about eating. Cunning and clever, her essays were equally about food and life. Even a passage about cauliflower from more than 70 years ago may still stop your reading in your tracks and send you off to the kitchen. She wrote, “There in Dijon, the cauliflowers were very small and succulent, grown in that ancient soil. I separated the flowerlets and dropped them in boiling water for just a few min- utes. Then I drained them and put them in a wide shallow casserole, and covered them with heavy cream, and a thick sprinkling of freshly grated Gruyere, the nice rubbery kind that didn’t come from Switzerland at all, but from the Jura. It was called râpé in the market, and was grated while you watched, in a soft cloudy pile, onto your piece of paper” (Fisher, 1943, p. 441). Most people eat cauliflower only because it is healthy, but that description may cause you to look up a recipe for a cauliflower casserole (Figure 15.11). What we can see in Fisher is the eloquent way in which she expresses the pleasure and joy food brings to us. Pick your favorite food—a slice of pepperoni pizza, poutine, sushi, Belgian chocolate, hot and sour soup, chana masala, or a breakfast omelet loaded with cheese and mushrooms—many foods bring us distinct joy. In this section, we discuss both the anatomy of the receptors that transduce chemicals in the mouth into taste sensations and the neural circuitry in the brain responsible for the sense of taste. We also discuss the term flavor (a technical term here), which is used to describe the combination of taste and olfactory sensations into a combined perception, which is often what we are appreciating when enjoying good food. Like the sense of smell, the sense of taste is a chemical detection system that prob- ably evolved to help us sort edible foodstuffs from toxins. We want to bring into our bodies only foods that are not toxic. So we have evolved to enjoy foods that bring nour- ishment and reject foods that do not. Our tongues therefore serve both as a gateway to nutrition and to warn us of toxic chemicals. The fact that food brings us pleasure reinforces the need to eat and supply our body with nutrition. There are basic similarities between the two chemical senses of olfaction (smell) and gustation (taste). Both respond to molecules, but for the sense of taste, we respond to molecules called tastants, which dissolve when in contact with our saliva. Like odorants, tastants initiate a transduction response in the receptors in our tongues. Taste, therefore, is the perception of the transduction of tastants along the surface of the tongue. There are also taste buds in other areas of the mouth that detect taste. However, what we typically crave when we think of our favorite food is its flavor. Flavor refers to the combined sensory experience of a food, which combines its taste, its odor, and its effect on the trigeminal nerve. If we think of the flavor of coffee, the smell of coffee is more instrumental to the flavor of this beverage than is its taste. If we think of the flavor of pizza, the smell of that pizza slice is as critical to flavor as is its taste. Tastes can be broken down into five basic tastes, which correspond to specific FIGURE 15.11 Cauliflower Casserole Shutterstock.com /m inadezhda Tastants: molecules recognized by taste receptors that induce responses in taste receptors on the tongue Taste: the perception of the transduction of tastants along the surface of the tongue Flavor: the combined sensory experience of a food, which combines its taste, its odor, its effect on the trigeminal nerve, and even visual experience 448 Sensation and Perception receptors along the surface of the tongue. The basic tastes are sweet, salty, sour, bitter, and umami (savory) (Figure 15.12). Sweet, salty, and umami tastes alert us to foods with needed nutrients. Sweet means sugar and therefore car- bohydrates, which we need to survive. Salty foods obvi- ously contain salt, which our bodies require as well, and umami (a Japanese word meaning “savory”) is associ- ated with the taste of proteins, which our bodies need to break down into amino acids in order to build our own proteins. In contrast, though we may like sour and bitter foods in small doses, it is likely that sour and bitter tastes evolved as warnings about foods to avoid because they may be inedible or toxic. Young children tend to avoid very sour and bitter foods because of this innate reaction to them. This is why young children invariably reject bitter foods such as broccoli or coffee. Try giving a 3-year- old an endive and kale salad, and you will have a tantrum on your hands. We slowly acquire the pleasure of bitter foods, usually as we combine them with sweet, salty, and savory tastes. We also find that many of the bitter foods others eat do not make us sick after all and thus acquire a taste for them. Even so, an adult eating an endive and kale salad probably puts a dressing on that salad that adds sweetness, saltiness, or umami to it. Sugars are the tastant molecules that give food its sweet taste. These include sucrose (the sugar commonly used as table sugar), fructose (the sugar found in fruits and used in many processed foods), and glucose (which is used as an energy source in the body). Artificial sweeteners, such as aspartame, are built from amino acids but mimic the actions of sugars along the tongue. The salty taste is elicited by foods that contain the chemical associated with common table salt, sodium chloride (NaCl). The sodium part of salt is necessary for a number of bodily functions, so a desire for salty foods pushes us to seek out and eat salty foods. Sour tastes are caused by acids and may be pleasurable at low concentrations. Many of us like sour foods such as grapefruits and other citrus fruits, pickles, kimchi, and yogurt. Bitter tastes are elicited by a number of different kinds of molecules. Most bitter tastes arise from plant substances that are used to discourage animals from eating their leaves and other parts of the plants. Some salad greens, such as endive and kale, have a bitter taste because those plants evolved to deter would-be eaters by tasting bad. Ironically, these plants now thrive because humans like their bitter taste. Umami is now considered a basic taste, though its inclusion in that group is rela- tively recent (Yasuo, Kusuhara, Yasumatsu, & Ninomiya, 2008). There is evidence that umami receptors exist on the tongue that are anatomically distinct from other recep- tors. These are activated by amino acids found in foods such as meats and mushrooms, as well as by the chemical MSG (monosodium glutamate). Umami signals the presence of amino acids, which we need in our diet to synthesize proteins. Traditionally, these amino acids come from meat or dairy products, although many plant foods also contain amino acids. These plant-based amino acids, such as in tomatoes and mushrooms, also trigger umami responses. Anatomy of the Tongue and Taste Coding Tastants are detected by receptors, located mostly on the tongue but also on other surfaces within the mouth. Most taste buds are on the tongue, but about 33% of taste buds can be found on the epiglottis, the soft palate, and the upper esophagus. These receptors, then, as we have seen with other sensory systems, transmit a neural signal to Sweet Salty Sour Umami Bitter FIGURE 15.12 The Five Basic Tastes 449 Chapter 15: Olfaction and Taste the brain, which then interprets the signal and integrates it with other sensory informa- tion. So here’s the anatomy that supports taste transduction, and then we will follow the neural signal. The human mouth contains approximately 10,000 taste buds (Bartoshuk, 1971), most of which are found on the tongue (Figure 15.13). Taste buds are small structures located along the surface of the tongue or mouth that contain the receptor cells. Taste buds are found within papillae, which line the surface of the tongue and mouth. There are four different kinds of papillae, three of which contain the taste buds. The four dif- ferent kinds of papillae are referred to as fungiform papillae, foliate papillae, circum­ vallate papillae, and filiform papillae. The fungiform papillae are located mostly along the edges and top of the tongue. The foliate papillae are found along the side of the tongue. The circumvallate papillae are found along the very back of the tongue, aligned seemingly along a row. The filiform papillae are found all over the tongue. But rather than taste buds, these papillae contain somatosensory receptors. See Figure 15.14 for the anatomy of the tongue. Each taste bud contains anywhere from 40 to just more than 100 taste receptor cells, which are elongated neurons with cilia at the end. Taste receptor cells are the cells within the taste buds that transduce tastants into a neural signal. As with the olfactory receptors, taste receptor cells die off after about a week and are replaced by new taste receptor cells that develop within the taste bud. Tastants come into con- tact with the surface of the tongue, where they are felt by the cilia and transduction may occur. Our tongues contain two kinds of taste receptor cells. One type is simply called receptor cells. Receptor cells transduce sweet tastes, umami tastes, and bitter tastes. Anatomically distinct presynaptic cells are also receptor cells, but they transduce salty FIGURE 15.13 An Actual Human Tongue The little bumps are the papillae, which contain the taste buds. Shutterstock.com / Raw pixel.com Foliate Circumvallate Filiform Fungiform papilla Fungiform Taste pore Taste bud Taste cell Nerve fiber Receptor sites on tip of taste cell Bitter Sweet Sour Salt Na+ H+ Taste buds FIGURE 15.14 Anatomy of the Tongue The human tongue contains four kinds of papillae, three of which contain taste buds. Circumvallate papillae are located in a line along the back of the tongue. Foliate papillae are located along the sides of the tongue. Fungiform papillae are widely distributed but more concentrated toward the front of the tongue. Taste buds: small structures located along the surface of the tongue or mouth that contain the receptor cells Papillae: small structures that contain the taste buds Fungiform papillae: located mostly along the edges and top of the tongue; they respond to all five basic tastes Foliate papillae: found along the side of the tongue; they respond to all five basic tastes Circumvallate papillae: found along the very back of the tongue in a virtual row; they respond to all five basic tastes Filiform papillae: found all over the tongue; they contain somatosensory receptors rather than taste buds, so that they feel food rather than taste it Taste receptor cells: cells within the taste buds that transduce tastants into a neural signal Receptor cells: taste receptor cells that transduce sweet tastes, umami tastes, and bitter tastes Presynaptic cells: taste receptor cells that transduce salty and sour tastes 450 Sensation and Perception and sour tastes. In receptor cells, the cilia contain receptors that respond to one and only one kind of taste, but unlike earlier conceptions, the different tastes are mixed across the surface of the tongue (Chandrashekar, Hoon, Ryba, & Zuker, 2006). Each presynaptic cell can transduce both salty and sour tastes. Presynaptic cells are connected to cranial nerve fibers, which transmit the signal to the brain. Receptor cells appear to have specific responses they transmit to presynaptic cells, which then transmit the signal to the brain, although this idea is still quite controversial (Yoshida & Ninomiya, 2010). Information leaves the taste buds through the 7th, 9th, and 10th cranial nerves and heads to the brain. It then synapses in the nucleus of the solitary tract in the medulla and then travels to the ventral posterior medial nucleus of the thalamus before heading to the cortex. The first area of the cortex that receives input from the taste system is the anterior insular cortex (part of the area known as the insula), located in the frontal lobe. The insular cortex has a variety of functions in addition to taste perception, but it is sometimes called the gustatory cortex for its functions in taste perception. The insular cortex also projects to the orbitofrontal cortex, where taste and olfaction mix to create flavor perception. The orbitofrontal cortex also has a number of other functions, including a role in emotion and emotional control. Figure 15.15 shows an anatomical drawing of the brain’s network for taste. Some interesting work has been done on how taste perception centers in the orbi- tofrontal cortex. When rats eat food, we can see activity in the anterior insular cortex, regardless of whether the rats are hungry or not. However, in the orbitofrontal cortex, there is activity only when the rats are hungry (Rolls, 2006). Thus, it is thought that the orbitofrontal cortex is largely responsible for the pleasure involved in the eating of food. This is compounded by the fact that the orbitofrontal cortex also responds to a number of other sensory modalities that process the quality of food. TEST YOUR KNOWLEDGE 1. What does the term umami mean? What chemicals elicit its perception? How does it differ from bitterness? 2. Describe the neural pathways from the tongue through higher order areas of the brain. Thalamus Nucleus of the solitary tract Chorda tympani (branch of the facial nerve, CN VII) Glossopharyngeal nerve (CN IX) Vagus nerve (CN X) Primary taste cortex (insula) Brain stem Tongue Nerve fibers innervating taste receptor cells Circumvallate papillae Foliate papillae Fungiform papillae Taste receptors: FIGURE 15.15 Pathways of Taste to the Brain Information from the tongue is sent through the cranial nerves to the nucleus of the solitary tract, then to the ventral posterior medial nucleus of the thalamus, then to the insula, and from there to the orbitofrontal cortex as well as areas in the limbic system. Anterior insular cortex (insula): a part of the frontal lobe that serves as the primary taste cortex 451 Chapter 15: Olfaction and Taste Taste and Flavor We have mentioned throughout this chapter how important the interaction is between our two chemical senses, taste and olfaction. Indeed, the concept of flavor refers to the combination of both of these senses and others in determining the basic pleasantness of foods. Thus, we define flavor as the total perceptual experience that occurs during eating, which combines not only taste and olfaction but also somatosensory expe- rience and visual experience. For example, think about drinking a cup of hot chocolate. Flavor is composed of olfactory sensations—the smells wafting up to the olfactory epithelium. It is also composed of the response along the tongue, in particular, the unique combination of sweet- ness and bitterness evoked by chocolate. Also important to flavor is the action of the somato- sensory system. In this case, the hotness of the hot chocolate is registered by thermoreceptors in the mouth and thus contributes to flavor as well. If there is a minty taste to your hot choco- late, the trigeminal nerve contributes to flavor as well. Finally, even vision and audition contribute to flavor. The sight of the chocolaty liquid with the marshmallows and cinnamon flakes on top is part of the delight. The sound of crunchiness may be an important part of flavor for some foods (though not hot chocolate). We eat potato chips because we crave salt, but the crunch they make in our mouth is important to flavor. Much of this integration takes place in the orbitofrontal lobe (Figure 15.16). Individual Differences in Taste Perception As with other sensory modalities, there are individual differences in our ability to taste. We probably notice these differences whenever we eat with friends and family. Some may require more salt to make their food taste the way they want it, and some may drown their food in chili pepper sauce. Some of these individual differences have a genetic basis. Just as most of us have a three-cone visual system (but some of us have a two-cone [or fewer] visual system), genetic differences affect taste perception as well. By far the most widespread of these genetic differences is differences in our ability to taste bitter foods. There is a particular gene, known as TAS2R38, that codes for bitter-taste receptors. The TAS2R38 gene comes in two forms, the PAV form and the AVI form. Most of us have the PAV form, which allows us to detect bitter tastes, especially those that derive from two chemicals known as phenylthiocarbamide and propylthiouracil. People with the PAV form of the gene who can detect these chemicals are called simply tasters. Most of us are just plain tasters. However, a minority of about 25% of people, who have the AVI form of the gene, are referred to as nontasters because they require much higher doses of these chemicals in order to detect a bitter taste (Bartoshuk, Duffy, & Miller, 1994). Nontasters will enjoy many bitter foods not because they sense the bitter taste but rather because other tastes are more salient to them (Figure 15.17). Hunger/thirst Vision Audition OLFACTION FLAVOR TASTE Onset/ aftertaste Intensity Quality Hedonics Pleasant Unpleasant Localization “Mouth feel,” including factors such as • Texture (crispiness, fattiness, grittiness, etc.) • Temperature (cold to hot) • Spiciness (as from chili peppers) • Coolness (as from mint) • Dryness (astringency, as from unripe fruit) Sweet Salty Umami Sour Bitter Mild sour Strong sour FIGURE 15.16 The Perception of Flavor Taste and olfaction, as well as vision, audition, and the somatosensory system, all contribute to our perception of flavor. Sweet, salty, and umami tastes alert us to foods with needed nutrients. Sweet means sugar and therefore carbohydrates, which we need to survive. Salty foods obviously contain salt, which our bodies require, and umami (a Japanese word meaning “savory”) is associated with the taste of proteins, which our bodies need to break down into amino acids in order to build our own proteins. Tasters: people who can detect bitter compounds Nontasters: people who cannot detect bitter compounds except at very high concentrations 452 Sensation and Perception There is also a class of people known as supertasters, who have the PAV form of the gene but who also have more fun- giform papillae on their tongues than normal tasters do, also from genetic causes. Because of the genetic origins of supertasters, there is a higher percent- age of supertasters among populations of non-European origins. Asians, for example, are more likely to be super- tasters than Europeans. In addition, there is a sex difference. Women are more likely than men to be supertasters. However, recent data suggest that fun- giform papillae density is not related to being classified as a supertaster (Garneau et al., 2014). Thus, the neural, but not genetic, basis of supertasters is now in some doubt. Because of their high concentration of fungiform papillae, supertasters are highly sensitive to bitter tastes and usually avoid them, as they find bitter tastes overwhelming. Thus, supertasters tend to avoid bitter beverages, including beer, green tea, and cof- fee, which contain many bitter compounds. In addition, supertasters may avoid bitter foods, such as Brussels sprouts, kale, cabbage, olives, some soy products, and some citrus fruit, such as grapefruits. Supertasters may also avoid spicy foods because the fungiform papillae are innervated by the nerve fibers that send information about burn and touch sensations on the tongue. Thus, supertasters may also avoid foods containing chili peppers (Delwiche, Buletic, & Breslin, 2001). To summarize, there are individual differences in the detection of tastes, some of which derive from genetic differences among individuals. Some research shows that these distinctions have health implications. Supertasters avoid some healthy foods because they just taste too bitter. In support of this view, Duffy, Lucchina, and Bartoshuk (2004) found that older supertasters had more colon polyps (a precursor to colon cancer) than did control participants because they avoided vegetables that lower the risk for these colon polyps. However, they also found that because supertasters avoid fatty foods as well, they show lower risk for cardiovascular disease. The Wonderful World of Chili Peppers In Chapter 2, we introduced the effect of chili peppers on taste perception as an exam- ple of how psychophysics works in the real world. We discussed the Scoville scale and how it measures our perception of capsaicin, the active ingredient in chili peppers that causes them to be experienced as piquant or “spicy hot.” In this section, we briefly describe the process by which we taste the capsaicin in chili peppers (Figure 15.18). Capsaicin content varies in different peppers, from relatively small amounts in a banana pepper to enough to cause a “heat” response in a habanero pepper to the specially bred chili peppers of today with very high capsaicin content, such as the Carolina Reaper. The painful or burning sensation caused by capsaicin in chili peppers is caused by its activation of activity in the trigeminal nerve, which codes for heat in the mouth. Capsaicin also stimulates taste receptors in the fungiform papillae, which contribute to the sensation of heat on the tongue. It is the action along the fungiform papillae that make supertasters more sensitive to capsaicin than are normal tasters. Despite the feeling of pain that eating a very hot chili pepper may bring, it causes no actual physical FIGURE 15.17 Receptors on the Tongues of (a) Nontasters and (b) Supertasters The tips of the tongues of a nontaster (a) and a supertaster (b) (pink = fungiform papillae). Supertasters: people who are extremely sensitive to bitter tastes; they usually do not like foods with many bitter compounds FIGURE 15.18 The Joys of Chili Peppers © iS to ck ph ot o. co m /G M Vo zd 453 Chapter 15: Olfaction and Taste damage. There may be slight swelling or inflammation, but this is caused by the actions of the nerves themselves, not the capsaicin. Interestingly, exposure to capsaicin causes the receptors to desensitize to the chemical. Thus, if you accidentally eat a hot chili pepper and wait until the burn subsides, you will find yourself curiously able to handle another one (and impress your friends). This desensitization ironically allows capsaicin to be used as a painkiller in some medical applications. Development of Taste Perception Infants are seemingly born with an attraction to foods that are sugary and foods that are salty (Herz, 2007). As any parent knows, sweet foods will elicit smiles in young infants (Figure 15.19). Virtually every family has a photo- graph of a happy baby whose face is covered in chocolate cake. However, many of our taste preferences are condi- tioned responses that develop as a function of experience and reinforcement. For example, most infants and young children avoid coffee, as it is too bitter for their tastes. We acquire a liking for the taste of coffee as we associate it with the sugar we put in it and the milk we put in it, as well as the wakefulness that the caffeine in coffee causes. Thus, by the time we reach adulthood, many of us enjoy both the aroma and the taste of coffee, even when not combined with milk or sugar. Similarly, most young chil- dren will reject alcoholic beverages unless they are paired with large concentrations of sugar. This is true for some adults as well. Beverages such as rum and tequila are popular because they are very sweet as well as highly alcoholic. And, of course, few parents would subject their infants to Scotch bonnet peppers. This taste preference develops later in life. Some research shows how a person’s early environment can affect taste perception. For example, an early deficiency in salt intake can cause later cravings for and higher intake of salts (Stein et al., 1996). Even when pregnant mothers do not receive adequate salt in their diet, their offspring will show later salt cravings (Crystal & Bernstein, 1995). Similarly, our desire for fatty foods may also be influenced by early developmen- tal experiences. TEST YOUR KNOWLEDGE 1. How do taste perceptions interact with olfaction to create flavor? 2. What are the differences between tasters, nontasters, and supertasters? FIGURE 15.19 Young Children Crave Sugary Foods Shutterstock.com /aaltair EXPLORATION: Anosmia Carl is a young man in his late teens. He attends the local community college, where he made the dean’s list in his first year. Carl is athletic, outgoing, and a dog enthusi- ast. He never misses walking his family’s four dogs every morning. Carl’s father was a champion professional boxer and has trained Carl to box. Although Carl’s career ambitions lie elsewhere, he frequently spars at the local boxing gym with other young men. While sparring, Carl was punched in the nose, breaking the cartilage and forc- ing some of the cartilage up through his olfactory tract, fracturing the cribriform plate and shearing the olfactory nerve. Since then, Carl has had a condition called anosmia, 454 Sensation and Perception the inability to perceive odor because of malfunctioning olfactory perception. Carl has complained that food does not taste as good, and he has lost a bit of weight. But he now seeks out more high-calorie food to compensate for his loss of appetite. Interestingly, he regrets not being able to smell his dogs, but in general, he has adjusted to his condition. Anosmia such as this may be only temporary, or it may be permanent, depending on the damage to the cribriform plate and whether the olfactory nerve was just damaged or completely cut. In some cases, as long as the olfactory nerve is not severed, olfaction may eventually come back. In addition to head trauma, anosmia may be caused by inflammation in the nasal cavities, blockage of the nasal passageways (in which case the anosmia may be tempo- rary), or damage to the areas of the brain responsible for olfactory perception (in which case the anosmia is likely permanent). Temporary anosmia may also develop in depressed patients receiving electroconvulsive shock therapy. Anosmia may occur in progressive neurological disorders such as Parkinson’s disease and Alzheimer’s dis- ease (Attems, Walker, & Jellinger, 2014). Indeed, patients’ awareness of their anosmia may be one of the early clues that lead doctors to make these more serious diagnoses. There are also people who are born with an extremely rare condition known as congenital anosmia. Such individuals do not develop a normal sense of smell or any sense of smell at all (Frasnelli, Fark, Lehmann, Gerber, & Hummel, 2013). Most people with congenital anosmia may not realize it until well into adulthood, as in most situations, they may compensate for it with their other senses. Anosmia may result also from traumatic brain injury to areas of the brain receiving input from the olfactory bulb. In particular, given its location near the front of the skull and behind the eyes, the orbitofrontal cortex may often be damaged in accidents, such as car crashes. Damage to the orbitofrontal cortex can result in anosmic symptoms (Caminiti, Ciurleo, Bramanti, & Marino, 2013). Until recently, the different causes of anosmia were grouped together, but researchers have begun to notice differences in the etiology resulting from damage to different sections along the olfactory tract. Think back to our discussion of different forms of visual impairment. Damage to the eyes can cause blindness. However, damage to the visual areas of the brain is more complex. Damage to V1 causes blindness (except for areas of blindsight), but damage to higher areas of the extra- striate cortex can cause complex agnosias, such as the inability to identify objects, even though a patient can see them and describe them. A similar situation exists in the olfactory system, though it is just being acknowledged and investigated. Shearing of the olfactory nerve causes a loss of olfactory sensation. People who, like Carl, have severed olfactory nerves can no longer detect odors. Thus, they are unable to recognize them, but it is likely that if they possessed olfactory imagery, it might still exist. However, recent research on anosmia in people with orbitofrontal damage suggests that these people can still detect odors, but they are extremely impaired in recognizing them rela- tive to controls (Caminiti et al., 2013). Thus, damage to the orbitofrontal cortex may be more similar to a form of olfactory agnosia than anosmia. That is, damage to the nose itself and the olfactory bulb pro- duces deficits in the detection of odors, whereas damage to the piriform cortex and orbitofrontal cortex produces deficits in odor recognition and discrimination. For exam- ple, Tranel and Welsh-Bohmer (2012) identified a patient with extensive damage to his limbic system and orbito- frontal cortex and found intact detection of odors but severely impaired naming, recognition, and discrimination of odors. Anosmia may also occur in the normal process of aging. It is estimated that by the age of 80 years, nearly half of the population has experienced some deficit in their sense of smell (Herz, 2007). Because this loss is gradual, many older adults may not be aware that their senses of smell have diminished. Although we cannot compen- sate for olfactory loss (i.e., there is no such invention yet as a smelling aid), awareness of the effects of anosmia might help some older adults cope with the problems caused by anosmia. For example, if the sense of smell is compromised, food may not be as appealing, leading to declines in caloric intake and proper nutrition. In fact, Herz reported that older adults have fewer “cravings” for specific foods than do younger adults, which may be due partially to reduced olfactory abilities. Older adults may also add more salt to their food, which in some cases can result in increased blood pressure. Adding salt may com- pensate for the loss of flavor due to lost olfactory sensitiv- ity, but it may compromise their health. We can experience something akin to this form of anosmia when we have a very stuffed nose. When we have such a cold, food does not have the same flavor, because we are not getting the same olfactory input. With a cold, you may find that you avoid foods that rely on olfaction for their flavor, such as coffee, and prefer foods with heavy salt content (e.g., chicken soup). 455 Chapter 15: Olfaction and Taste Anosmia is also linked with depression. The higher areas of the olfactory system are closely linked to the limbic sys- tem, which is involved in the regulation of emotions. Many brain regions associated with olfaction, such as the orbi- tofrontal cortex, are considered part of the limbic system. Thus, the loss of smell may affect emotion (Herz, 2007). It is likely that there is a causal direction between anos- mia and depression because brain areas associated with olfaction and emotion are connected, such that damage to the olfactory regions also impairs emotional responding. Some studies suggest that depression may develop gradu- ally after the onset of anosmia. There also exists a neurological condition called phan- tosmia (Landis, Croy, & Haehner, 2012). Phantosmia is a condition in which people smell odors that are not physically present. In this sense, phantosmia is essen- tially an olfactory hallucination, which can be quite dis- tracting to patients who have it. It differs from olfactory imagery because the person is not in control of olfactory experience, nor is he aware that the experience is differ- ent from real odor perception. Phantosmic smells occur without any conscious control over them. Phantosmia may occur during early Parkinson’s disease and in other neurological disorders. Phantosmia may also occur as a precursor to an epileptic seizure. Similarly, phantosmia may also occur as part of an “aura” just before a patient experiences a migraine. Fortunately, though, most bouts of phantosmia tend to dissipate over time (Landis et al., 2012). Thus, in most cases, phantosmia is not permanent (though if the smell is of chocolate, one might want it to last longer). Phantosmia is an interesting area, but it is one that awaits further study. And in case you were wonder- ing, ageusia is the term that describes a person who has lost the ability to taste. It is an extremely rare condition, especially for all tastes. Phantosmia: hallucinatory perception of odors Ageusia: loss of the ability to taste APPLICATION: Artificial Sweeteners and Taste Perception As we have discussed previously, most people crave sweet foods. In past times, this served humans well, as calories were in short supply and sweet foods contained lots of cal- ories. In today’s world (at least in the Western world), cal- ories are abundant, and eating too many sweet foods is a danger instead of being helpful. High-calorie foods, espe- cially sweet foods, can lead to obesity, type 2 diabetes, and various forms of heart disease (Frazier, 2018). As such, many people now opt for diet foods with artificial sweet- eners (Figure 15.20). Artificial sweeteners are intended to impart fewer calories but provide the same or similar taste of sweetness to people who use them. Our goal here is not to debate their value for health but to understand how artificial sweeteners, which are usually amino acid based rather than sugar based, stimulate receptors on the tongue and how they differ in perception from actual sugars (see Swithers, 2016, for an empirical-based discussion of the healthfulness of artificial sweeteners). We define artificial sweeteners as sugar substitutes that seek to duplicate the effect of sugar on our sense of taste, but with fewer calories. Most artificial sweeteners work because they hyperstimulate the taste bud receptor cells that are responsible for sweet tastes. Artificial sweeteners stimulate receptor proteins on the outer tips of the taste cells that respond to sweetness and therefore induce an experience of sweetness. That is, artificial sweeteners, such as saccharin and aspartame, can be used as sweeteners because they elicit a sweet response from our taste buds but do so with far less of the chemical than standard nat- ural sugars, such as sucrose or fructose. Because saccharin can stimulate sweet receptors at lower levels, less of it is needed, and, therefore, fewer calories are consumed by the individual. Indeed, by weight, saccharin is many times more powerful at eliciting sweet responses than is sucrose. Some estimates hold that saccharin is 300 times more suc- cessful at stimulating sweet receptors than actual sugars. If this is so, it is natural to ask why is it that many of us Artificial sweeteners: sugar substitutes that seek to duplicate the effect of sugar on our sense of taste, but with fewer calories 456 Sensation and Perception still prefer the taste of sugars to that of artificial sweeten- ers? Why do sugared sodas taste so much better than diet sodas to so many of us? As discussed earlier, the neural pathway for taste takes information to the primary taste cortex in the anterior insula. The anterior insular cortex, also known as the primary gustatory cortex, projects to the orbitofrontal cortex, where taste and olfaction mix to create flavor perception. However, the primary taste cortex also pro- jects to areas associated with reward pathways, such as the amygdala and the caudate nucleus. These areas are related to the release of dopamine, a neurotransmitter associated with reward and pleasure. It turns out that natural sugars promote activity in the amygdala and the caudate nucleus, but artificial sweeteners do not (Frank et al., 2008). Another reason that natural sugars taste bet- ter than artificial sweeteners is that many artificial sweet- eners also have a bitter taste, which interferes with the experience of the sweetness (Rodrigues et al., 2016). For this reason, some research is directed at eliminating the bitterness from artificial sweeteners so that the sweetness can be more effective. Frank et al. (2008) used magnetic resonance imaging (MRI) technology to compare the effects of sugar and artificial sweeteners on various regions of the brain. Participants were given drinks with either sugar or arti- ficial sweetener (sucralose) and asked to rate them on a scale in which 1 represented “did not taste good” and 9 represented “extremely enjoyable.” Frank et al. found that both the sugar and the artificial sweeter activated the pri- mary taste area in the insula, but only the sugar activated the reward areas in the midbrain and caudate nucleus. This suggests, as we know from earlier discussion, that there is more to taste than just the response on the tongue. It is likely that satiation mechanisms induce a feedback loop, which results in the high-calorie sugars still tasting better, even though the artificial sweeteners may be causing the same or bigger response in the taste buds. However, some research suggests that consistent use of artificial sweeten- ers may eventually result in bigger responses in the mid- brain, perhaps as people get used to the different internal feedback from the artificial sweetener (Green & Murphy, 2012). Other factors may also play a role. For example, Wang, Duncan, and Dietrich (2016) varied the mineral content of water, that is, the presence of chemicals such as iron or calcium in drinking water. Natural sugars were judged as being more enjoyable when mixed with these chemicals than artificial sweeteners. However, the data are not always that sugars are more pleasant than artificial sweeteners. Delogu et al. (2016) found that participants in a blind taste test judged as more pleasant carbonated beverages sweetened with artificial sweeteners. However, it was unclear how this study accounted for the amount of sweetener in each beverage because this study used com- mercially available drinks. Nonetheless, it does suggest, that in some circumstances, the artificial sweeteners may be considered sweeter than natural sugars. There are also a few natural nonsugar-based sweeteners, that is, natural products that serve as “artificial sweet- eners.” The most famous of these is the miracle fruit, or Synsepalum dulcificum, as it is known in scientific lingo. The miracle fruit itself has a low sugar content, but it also has a chemical (called, aptly, miraculin) that activates the FIGURE 15.20 Artificial Sweeteners Artificial sweeteners are designed to provide eaters with the sweet taste they crave, but without an abundance of calories. Artificial sweeteners are chemicals that highly stimulate receptors on the tongue responsible for sweet tastes. © iS to ck ph ot o. co m /N oD er og (a) Shutterstock.com /SpeedKingz (b) Chapter 15: Olfaction and Taste 457 taste bud receptors responsible for sweet taste (Misaka, 2013). In particular, when it is eaten in combination with sour foods, such as lemons or oranges, it causes the sour taste of these foods to taste sweet because of this activation of the sweet taste receptors. The effect of the miraculin lasts up to about 30 minutes (see Rodrigues et al., 2016). For example, Rodrigues et al. reported that unsweetened lemonade tasted sweet to participants when miracle-fruit juice was added. Currently, miraculin is not allowed to be used as a sweetener in American markets, but it is avail- able as a supplement in some European countries. Artificial sweeteners are somewhat controversial. There are those who think that the use of artificial sweeteners will help people overcome obesity and diabetes by provid- ing a low-calorie alternative to obtaining sweetness (but see Swithers, 2016, for a dissenting view). High-impact artificial sweeteners may also be useful for people expe- riencing loss of appetite for one reason or another (e.g., chemotherapy). Thus, understanding how and why arti- ficial sweeteners produce sweetness and how they can be designed to mimic the taste of natural sugars is an inter- esting applied issue in perception research. CHAPTER SUMMARY 15.1 Examine the role of the olfactory system in detecting harmful and helpful substances. The chemical senses allow us to detect molecules either in the air or in our mouths. Olfaction is the sense of smell, whereas gustation is the sense of taste. Olfaction detects odorants, volatile chemicals in the air that enter our nose. Odors are the perceptual experience of odorants, which are airborne chemical stimuli. Odorants enter our noses, where they head up the two nostrils, which are separated from each other by the nasal septum. Turbinates then disperse air within the nasal cavity. The odorants rise to the olfactory cleft, the channel at the back of the nasal cav- ity that funnels air up toward the olfactory epithelium. The olfactory epithelium is a mucous membrane inside each nostril of the nose that contains the receptor cells for the olfactory system. The olfactory receptor neurons are the receptor cells located in the olfactory epithelium. When the olfactory receptor neurons detect specific chemicals in the air, they transduce them into a neural signal. The signal is then sent to brain structures associated with the olfactory system. The olfactory receptor neurons are replaced fre- quently by the generation of new ones from the basal cells. We have about 350 different kinds of olfactory receptors, depending on how many genes we have activated to code for different forms of olfactory receptors. Just internal to the nose is the trigeminal nerve, also known as the fifth cra- nial nerve, which is associated with the feel of odorants, such as the burn of chili peppers. 15.2 Describe the anatomical and physiological bases for olfaction. The olfactory receptor neurons project axons through lit- tle holes through the cribriform plate, which travel to the olfactory bulb. If the olfactory tract is severed, anosmia, a deficit in olfactory perception, may result. The olfactory nerve starts with olfactory receptor neurons that leave the nose and enter the olfactory bulb. The olfactory bulb is a part of the brain just behind the nose. It is the first place in the brain where olfactory information is processed. The olfactory tract leaves the olfactory bulb to other regions of the brain, including the amygdala, entorhinal cortex, and piriform cortex. The piriform cortex is considered the pri- mary olfactory cortex. The piriform cortex is divided into two sections. The anterior piriform cortex is located in the front portion of the piriform cortex and is associated with representing the chemical structure of odorants. The pos- terior piriform cortex is located in the back portion of the piriform cortex and is associated with an odor’s quality, regardless of its chemical composition. The piriform cor- tex projects to the orbitofrontal cortex. The orbitofrontal cortex is critical in the emotional experience of odors and integrating olfaction and taste perception, among other functions. 15.3 Appraise the nature of olfactory perception, olfactory imagery, and olfactory illusions. Detection thresholds vary across odorants, though in gen- eral it does not take too many odorants to elicit an odor. We are also good at discriminating odors, that is, telling one from another. However, we appear not to be good at nam- ing odors. The tip-of-the-nose phenomenon occurs when a person is familiar with an odor but cannot recall its name, despite feeling as if she can. This experience of finding an odor familiar but being unable to name it is quite common. Most individuals do not have olfactory imagery, although a minority of people do. Those who do have olfactory imag- ery show activity in the piriform cortex when engaging in Sensation and Perception458 olfactory imagery. Olfactory illusions occur when our per- ceptions of odorants change depending on the external conditions. In one illusion, olfactory rivalry, one odorant is presented to one nostril, and a different odor is presented to the other nostril. In olfactory rivalry, we experience only one odor at a time, even though both are present. 15.4 Explain the anatomical and physiological bases for taste perception. Taste perception (gustation) is also a chemical system designed to make nutritious foods appealing and to make us withdraw from potentially toxic foods. The term tastant refers to molecules recognized by taste receptors that induce responses in taste receptors on the tongue. Taste is the perception of the transduction of tastants along the surface of the tongue, whereas flavor is the combined sensory experience of a food, which combines its taste, its odor, and its effect on the trigeminal nerve. Sweet, salty, and umami tastes alert us to foods with needed nutrients. Sour and bitter tastes evolved as warnings about foods to avoid because they may be inedible or toxic. However, in many cultural contexts, we have learned to enjoy sour and bitter foods as well. Taste buds are located along the sur- face of the tongue or mouth and contain the receptor cells that transduce the chemicals in foods into neural signals. The papillae are small structures that house the taste buds. There are four different kinds of papillae, distinguishable by their shapes, the kinds of taste buds they contain, and where along the tongue they are. Taste receptor cells are the cells within the taste buds that transduce tastants into a neural signal. There are two kinds of taste receptor cells. Receptor cells transduce sweet tastes, umami tastes, and bitter tastes, whereas presynaptic cells transduce salty and sour tastes. Information leaves the taste buds through the 7th, 9th, and 10th cranial nerves and then synapses in the nucleus of the solitary tract in the medulla and then travels to the ventral posterior medial nucleus of the thal- amus. The first area of the cortex that receives input from the taste system is the anterior insular cortex located in the frontal lobe. The anterior insular cortex is known as the primary gustatory cortex because of its functions in taste perception. The anterior insular cortex also projects to the orbitofrontal cortex, where taste and olfaction mix to create flavor perception. There are individual differences in taste perception, some of which are genetic in origin. Much research has inves- tigated the genes that code for the perception of bitter tastes. People can be divided into normal tasters, who can detect bitter compounds; nontasters, who cannot detect bitter compounds except at very high concentra- tions; and supertasters, who are extremely sensitive to and repelled by bitter tastes. Anosmia is a deficit in the detection or perception of odors. This condition can occur from damage to both the nasal pathway and the brain regions involved in olfaction. Phantosmia is the halluci- natory perception of odors. Ageusia is a loss of the ability to taste. Artificial sweeteners are sugar substitutes that attempt to duplicate the effect of sugar on our sense of taste, but with fewer calories. They do so by stimulating the same taste receptors on the tongue as natural sugars. However, the complexity of taste perception is such that people still prefer natural sugars to artificial sweeteners. REVIEW QUESTIONS 1. What are odorants? How do they differ from other molecules? Why do we have chemical detection systems? 2. What is the trigeminal nerve? How does it interact with olfaction and gustation? What basic sensations does it detect? 3. Describe the pathway from when an odorant enters the nose to when it is processed in the frontal lobes of the brain. Describe each step along the pathway. 4. What is the piriform cortex? What are the differ- ent functions of the anterior and posterior piriform cortices? 5. What is an olfactory illusion? Describe two olfac- tory illusions and what they tell us about olfaction in general. 6. Describe the gustatory pathway from when food is felt on the tongue to when it is processed in the fron- tal lobes of the brain. Describe each step along the pathway. 7. What is the difference between taste perception and flavor perception? What area of the brain seems to be responsible for flavor perception? 8. What are the five basic tastes? How are they pro- cessed on the tongue? Why is information about each basic taste essential for survival? Chapter 15: Olfaction and Taste 459 9. Describe the differences in genetics, anatomy, and behavior among tasters, nontasters, and supertasters. 10. What is anosmia? What causes anosmia to occur? How does anosmia resulting from damage to the nose differ from anosmia caused by damage to the piriform cortex? What is phantosmia? How does it differ from olfactory illusions? PONDER FURTHER 1. It is said that people can smell fear. How could this be? What would it take in order for human beings to realistically detect fear when other people experi- ence it? 2. Given that detecting poisons and seeking out calories and nutrients are such fundamental aspects of the sense of taste, how can it be that there is such amazing cultural variety in what humans eat across the world? KEY TERMS Ageusia, 455 Amygdala, 441 Anosmia (smell blindness), 440 Anterior insular cortex (insula), 450 Anterior piriform cortex, 442 Artificial sweeteners, 455 Basal cells, 438 Circumvallate papillae, 449 Cribriform plate, 440 Entorhinal cortex, 441 Filiform papillae, 449 Flavor, 447 Foliate papillae, 449 Fungiform papillae, 449 Glomeruli, 441 Gustation, 436 Macrosmatic, 439 Microsmatic, 439 Mitral cells, 441 Nasal septum, 437 Nontasters, 451 Odorants, 436 Odors, 436 Olfaction, 436 Olfactory bulb, 441 Olfactory cleft, 438 Olfactory epithelium, 438 Olfactory nerve (first cranial nerve), 441 Olfactory receptor neurons, 438 Olfactory tract, 441 Orbitofrontal cortex, 443 Papillae, 449 Phantosmia, 455 Piriform cortex, 441 Posterior piriform cortex, 442 Presynaptic cells, 449 Receptor cells, 449 Supertasters, 452 Supporting cells, 438 Tastants, 447 Taste, 447 Taste buds, 449 Taste receptor cells, 449 Tasters, 451 Tip-of-the-nose phenomenon, 444 Trigeminal nerve, 440 Tufted cells, 441 Turbinates, 438 Sensation and Perception460 Sharpen your skills with SAGE edge at edge.sagepub.com/schwartz2e SAGE edge offers a robust online environment featuring an impressive array of free tools and resources for review, study, and further exploration, keeping both instructors and students on the cutting edge of teaching and learning. Learning Objectives Digital Resources 15.1 Examine the role of the olfactory system in detecting harmful and helpful substances. Never Mind Eyesight, Your Nose Knows Much More Jennifer Pluznick: You Smell With Your Body Not Just Your Nose 15.2 Describe the anatomical and physiological bases for olfaction. Charting Plasticity in the Regenerating Maps of the Mammalian Olfactory Bulb Sniffing Out the Science of Smell Smelling in Stereo: Human Sense Detailed in Study 15.3 Appraise the nature of olfactory perception, olfactory imagery, and olfactory illusions. Nostril Advantage in Trigeminal/Olfactory Perception and Its Relation to Handedness Anosmia 15.4 Explain the anatomical and physiological bases for taste perception. Leaving a Flat Taste in Your Mouth: Task Load Reduces Taste Perception The Representation of Taste Quality in the Mammalian Nervous System Study: When Soda Fizzes, Your Tongue Tastes It Sushi Science: A 3-D View of the Body’s Wasabi Receptor Sriracha Chemistry: How Hot Sauces Perk Up Your Food and Your Mood This Is an Orchestra Under the Influence of Chili Peppers Glossary 461 GLOSSARY Absolute threshold: the smallest amount of a stimulus necessary to allow an observer to detect its presence Accommodation: the process of adjusting the lens of the eye so that both near and far objects can be seen clearly Accretion: the gradual reappearance of a moving object as it emerges from behind another object Acoustic reflex: a reflex that tightens the tensor tympani and the stapedius in response to chronic loud noise Acoustic shadow: the area on the side of the head opposite from the source of a sound in which the loudness of a sound is less because of blocked sound waves Action: any motor activity Active electroreception: the ability to generate electric fields and then detect disturbances or changes to those electric fields caused by external events Additive color mixing: the creation of a new color by a process that adds one set of wavelengths to another set of wavelengths A-delta fibers: myelinated nociceptors that conduct signals rapidly and respond to both heat and pressure Afferent fibers: neural fibers that carry sensory information to the central nervous system Affordance: information in the visual world that specifies how that information can be used Aftereffect: a sensory experience that occurs after prolonged experience of visual motion in one particular direction Afterimages: visual images that are seen after an actual visual stimulus has been removed Ageusia: loss of the ability to taste Agnosia: a deficit in some aspect of perception as a result of brain damage Akinetopsia (motion blindness): a rare condition in which an individual is unable to detect motion despite intact visual perception of stationary stimuli, caused by damage to area MT Ames room: a specially constructed room where two people of the same size standing in the two back corners will look very different in height Amplitude: the difference between maximum and minimum sound pressures Ampulla: the structure at the base of each semicircular canal that contains the crista Ampullae of Lorenzini: the organs that contain the hair cells that detect electric fields, used in passive electroreception Amusia: a condition in which brain damage interferes with the perception of music but does not interfere with other aspects of auditory processing Amygdala: an area of the brain in the limbic system, associated with the experience of emotion, particularly fear Analgesia: processes that act to reduce pain perception Anomalous trichromacy: a condition in which all three cone systems are intact, but one or more has an altered absorption pattern, leading to different metameric matches than in the most common type of trichromatic individuals Anosmia (smell blindness): the inability to smell, usually caused by cribriform plate damage Anterior chamber: the fluid-filled space between the cornea and the iris Anterior cingulate cortex: a region in the prefrontal lobe of the brain associated with the emotional experience of unpleasantness during pain perception Anterior insular cortex (insula): a part of the frontal lobe that serves as the primary taste cortex Anterior intraparietal (AIP) area: a region of the posterior parietal lobe involved in the act of grasping Anterior piriform cortex: a structure located in the front portion of the piriform cortex that is associated with representing the chemical structures of odorants Aphasia: an impairment in language production or comprehension brought about by neurological damage Apparent motion: the appearance of real motion from a sequence of still images Architectural acoustics: the study of how physical spaces such as concert halls affect how sounds are reflected in a room Artificial sweeteners: sugar substitutes that seek to duplicate the effect of sugar on our sense of taste, but with fewer calories Ascending series: a series in which a stimulus gets increasingly larger along a physical dimension Astigmatism: a condition that develops from an irregular shape of the cornea or the lens, which makes it impossible for the lens to accommodate a fully focused image Atmospheric perspective: a pictorial depth cue that arises from the fact that objects in the distance appear blurred and tinged with blue Attack: the beginning buildup of a note in music Attention: a set of processes that allow us to select or focus on some stimuli Attentional blink: the tendency to respond more slowly or not at all to the second appearance of a target in an RSVP task when the second target occurs within 500 ms of the first target Attentional capture: the process whereby a salient stimulus causes us to shift attention to that stimulus Audiogram: a graph that illustrates the thresholds for the frequencies as measured by the audiometer Audiologist: a trained professional who specializes in diagnosing hearing impairments Sensation and Perception462 Audiometer: a device that can present tones of different frequencies, from low in pitch to high in pitch, at different volumes from soft to loud Auditory core region: an area of the auditory cortex, consisting of the primary auditory cortex, the rostral core, and the rostrotemporal core Auditory cortex: the areas in the temporal cortex that process auditory stimuli Auditory scene analysis: the process of identifying specific sound-producing objects from a complex set of sounds from different objects at varying and overlapping frequencies Automaticity: refers to those processes that do not require attention Azimuth: the left-right or side-to-side aspect of sound localization Bálint’s syndrome: a neurological condition caused by damage to both left and right posterior parietal lobes Basal cells: cells that create olfactory receptor neurons Basilar membrane: the membrane that separates the tympanic canal from the middle canal; the organ of Corti lies on the basilar membrane Beat: spaced pulses that indicate if a musical piece is fast or slow Belt: a region of the auditory cortex that wraps around the auditory core regions Binocular cells: cells with two receptive fields, one for each eye; their main function is to match the images coming to each eye Binocular disparity: a binocular depth cue because our two eyes are in different locations in our head and therefore have slightly different views of the world Binocular rivalry: a phenomenon that occurs when a separate image is presented to each eye Biosonar: a process whereby animals emit sounds and then use comparisons of the emitted sounds and their returning echoes to sense the world around them Bistratified retinal ganglion cells (K cells): retinal ganglion cells that project to the koniocellular layer of the lateral geniculate nucleus; they represent 10% of ganglion cells, possess low sensitivity to light, and are sensitive to wavelength Blindsight: the presence of visual abilities even though a person experiences blindness because of damage to V1 Blobs: groups of neurons within V1 that are sensitive to color Bottom-up processing: a process whereby physical stimuli influence how we perceive them Brightness: the perceived intensity of the light present Broca’s aphasia: a form of aphasia resulting from damage to Broca’s area, causing a deficit in language production Broca’s area: an important area in the production of speech, located in the left frontal lobe Capsaicin: the active ingredient in chili peppers that provides the experience of hotness, piquancy, or spiciness Cataracts: a condition that results from a darkening of the lens Catch trial: a trial in which the stimulus is not presented Categorical perception: the perception of different acoustic stimuli as being identical phonemes up to a point at which perception flips to perceive another phoneme Center-surround receptive field: a receptive field in which the center of the receptive field responds opposite to how the surround of the receptive field responds; if the center responds with an increase of activity to light in its area, the surround responds with a decrease in activity to light in its area C-fibers: nonmyelinated nociceptors that are slower and respond to pressure, extreme degrees of either heat or cold, and toxic chemicals Change blindness: the difficulty we experience in detecting differences between two visual stimuli that are identical except for one or more changes to the image Characteristic frequency: the frequency to which any particular location along the basilar membrane responds best Chroma: the subjective quality of a pitch; we judge sounds an octave apart to be of the same chroma Ciliary muscles: the small muscles that change the curvature of the lens, allowing accommodation Circumvallate papillae: found along the very back of the tongue in a virtual row; they respond to all five basic tastes Coarticulation: the phenomenon in which one phoneme affects the acoustic properties of subsequent phonemes Cochlea: the snail-shaped structure of the inner ear that houses the hair cells that transduce sound into a neural signal Cochlear implants: devices that are designed to restore some hearing, typically of spoken voices, to deaf individuals; they stimulate the auditory nerve artificially with an electronic system, replacing the hair cells of the cochlea Cochlear nucleus: a structure in the brain stem that receives input from inner hair cells Cognitive impenetrability: perception is not affected by cognitive factors, only our reporting of perception is Cognitive penetration: the view that cognitive and emotional factors influence the phenomenology of perception Cold fibers: thermoreceptors that fire in response to colder (30 °C and below) temperatures as measured on the skin Color constancy: the ability to perceive the color of an object despite changes in the amount and nature of illumination Color deficiency: the condition of individuals who are missing one or more of their cone systems Color–music synesthesia: a form of synesthesia that occurs when particular pitches, notes, or chords elicit experiences of particular visual colors Glossary 463 Color-opponent cells: neurons that are excited by one color in the center and inhibited by another color in the surround, or neurons that are inhibited by one color in the center and excited by another color in the surround Complementary colors: colors on the opposite side of the color wheel that when added together in equal intensity give a white or gray or black Complex cells: neurons in V1 that respond optimally to stimuli with particular orientations; unlike simple cells, they respond to a variety of stimuli across different locations, particularly to moving stimuli Complex sound: a sound consisting of a mix of frequencies Compound eye: an eye that does not have a single entrance but is made up of many separate components called ommatidia Computational approach: an approach to the study of perception in which the necessary computations the brain would need to carry out to perceive the world are specified Conductive hearing loss: the inability of sound to be transmitted to the cochlea Cone of confusion: a region of positions in space in which sounds create the same interaural time and interaural level differences Cone-opponent cells: neurons that are excited by the input from one cone type in the center but inhibited by the input from another cone type in the surround Cones: photoreceptors in the fovea of the retina; they are responsible for color vision and our high visual acuity Congenital amusia: a condition in which people are inherently poor at music perception Conjugate gaze palsies: neurological disorders that affect the ability of the eyes to coordinate their movements; this inability to move together may affect eye movements in both vertical and horizontal directions Conjunction search: the search for a target in which the target is specified by a combination of features Consonance: the perception of pleasantness or harmony when two or more musical notes are played; that is, the notes fit with each other Consonants: speech sounds made with restricted airflow Constancy: the ability to perceive an object as the same under different conditions Constructivist approach: the idea that perceptions are constructed using information from our senses and cognitive processes Contralateral: literally, opposite (contra) side (lateral), meaning, in this context, that sensory information is on the side of the nervous system opposite the one from which it entered Contralateral organization: opposite-side organization; in the visual system, the nasal retina projects to the opposite side of the brain Contralateral representation of visual space: the arrangement whereby the left visual world goes to the right side of the brain, and the right visual world goes to the left side of the brain Convergence: the number of photoreceptors that connect to each ganglion cell; more convergence occurs for rods than for cones Cornea: the clear front surface of the eye that allows light in; also a major focusing element of the eye Corollary discharge theory: the theory that the feedback we get from our eye muscles as our eyes track an object is important to the perception of motion Correct rejection: in signal detection analysis, a correct rejection occurs when a nonsignal is dismissed as not present Correspondence problem (depth perception): the problem of determining which image in one eye matches the correct image in the other eye Correspondence problem (motion perception): how the visual system knows if an object seen at Time 1 is the same object at Time 2 Corresponding points: refers to a situation in which a point on the left retina and a point on the right retina would coincide if the two retinae were superimposed Cortical achromatopsia: loss of color vision due to damage to the occipital lobe Cortical magnification: the allocation of more space in the cortex to some sensory receptors than to others; the fovea has a larger cortical area than the periphery Covert attention: when your visual direction does not line up with your direction of gaze Cribriform plate: a perforated section of skull bone that separates the nose from the brain; axons from olfactory receptor neurons pass through to allow olfactory information to enter regions in the brain Crista: the structure in the ampulla of each semicircular canal that contains the receptors Criterion: a bias that can affect the rate of hits and false alarms Crossed disparity: the direction of disparity for objects closer to the viewer than the horopter (the image in the left eye is to the right of the image of the object in the right eye) Crossover point: the point at which a person changes from detecting to not detecting a stimulus or vice versa Cue approach to depth perception: the system whereby depth perception results from three sources of information, monocular cues to depth present in the image, binocular cues from the comparison of images in each eye, and cues from focusing the eyes, such as vergence and accommodation Cycle: in a sound wave, the amount of time between one peak of high pressure and the next d′ (d-prime): a mathematical measure of sensitivity Dark adaptation: the process in the visual system whereby its sensitivity to low light levels is increased Sensation and Perception464 Decay: how long the fundamental frequency and harmonics remain at their peak loudness until they start to disappear Decibel (dB): a physical unit that measures sound amplitude Deletion: the gradual occlusion of a moving object as it passes behind another object Dermis: the inner layer of the skin, which also houses touch receptors Descending series: a series in which a stimulus gets increasingly smaller along a physical dimension Deuteranopia: a lack of M-cones, leading to red–green deficiency; this trait is sex-linked and thus more common in men Difference threshold (JND): the smallest difference between two stimuli that can be reliably detected Diplopia: double images, or seeing two copies of the same image; usually results from the images of an object having too much disparity to lead to fusion Direct perception (Gibsonian approach): the approach to perception that claims that information in the sensory world is complex and abundant, and therefore the perceptual systems need only directly perceive such complexity Dishabituation: the process in which, after habituation has occurred, changing the stimulus causes the organism to respond again Dissonance: the perception of unpleasantness or disharmony when two or more musical notes do not fit together Distance: how far a sound is from the listener and whether it is in front of or behind the listener Divided attention: the process of attending to multiple sources of information Doctrine of specific nerve energies: the argument that it is the specific neurons activated that determine the particular type of experience Dorsal: in or toward the back of the body; in the head, it means at the top or toward the top Dorsal column–medial lemniscal pathway: a pathway for the mechanoreceptors (tactile perception) and proprioceptors (muscle position) that travels up the spinal column on the ipsilateral side and crosses to the contralateral side in the medulla Dorsal horn: an area of the spinal cord that receives input from nociceptors and feedback from the brain Dorsal pathway: starts with parasol retinal ganglion cells and continues through the visual cortex into the parietal lobe; often called the “where” pathway, as it codes for the locations of objects and their movement Dorsal root: the end of the spinal nerve where sensory information enters the spinal cord Dorsal root ganglion: a node on the spine where one finds nerve cells carrying signals from sensory organs toward the somatosensory areas of the brain Double-opponent cells: cells that have a center, which is excited by one color and inhibited by the other; in the surround, the pattern is reversed Dual target-cost: as the number of objects searched for increases, the likelihood of detecting one of those objects decreases Duplex theory of vision: the doctrine that there are functionally two distinct ways in which our eyes work, the photopic, associated with the cones, and the scotopic, associated with the rods Dynamics: relative loudness and how loudness changes across a composition Ecological approach to perception: another name for the direct perception view Edge completion: the perception of a physically absent but inferred edge, allowing us to complete the perception of a partially hidden object Edge detection: the process of distinguishing where one object ends and the next begins, making edges as clear as possible Electroencephalography (EEG): using electrodes to measure the electrical output of the brain by recording electric current at the scalp Electromagnetic energy: a form of energy that includes light that is simultaneously both a wave and a particle Electromagnetic spectrum: the complete range of wavelengths of light and other electromagnetic energy Electroreception: the ability to detect electric fields, seen in many species of fish Elevation: the up-down dimension of sound localization Endogenous opioids: chemicals produced by the body that reduce pain throughout the body Endolymph: fluid that fills the semicircular canals End-stopped neurons: neurons that respond to stimuli that end within the cell’s receptive field Entorhinal cortex: an area in the medial temporal lobe, associated with a number of memory functions Epidermis: the outer layer of the skin Equal-temperament scale: a tuning system in which the difference between each successive semitone is constant both in pitch and in frequency Eustachian tube: a thin tube that connects the middle ear with the pharynx and serves to equalize air pressure on either side of the eardrum Executive attention network: a system that focuses on attention by the inhibition of habitual responses and acts as the top-down control of attention; found in the frontal lobe Exploratory procedures: hand movements made in order to identify an object External auditory canal (external auditory meatus): the channel that conducts sound from the pinna to the tympanic membrane Extrastriate body area: an area within the inferotemporal cortex that is Glossary 465 activated when its cells view bodies or body parts but not faces Extrastriate cortex (secondary visual cortex): the collective term for visual areas in the occipital lobe other than V1 FAI mechanoreceptors: fast-adapting receptors, with Meissner corpuscle endings and small receptive fields, densely packed near the surface of the skin FAII mechanoreceptors: fast-adapting receptors with Pacinian corpuscle endings and large receptive fields, more widely distributed, deeper in the skin False alarm: in signal detection analysis, a false alarm is an error that occurs when a nonsignal is mistaken for a target signal Familiar size: the cue whereby knowing the retinal size of a familiar object at a familiar distance allows us to use that retinal size to infer distance Feature integration theory: a theory stipulating that some features can be processed in parallel and quickly prior to using attentional resources, whereas other visual characteristics require us to use attention and are done serially and therefore less quickly Feature search: the search for a target in which the target is specified by a single feature Field of view: the part of the world you can see without eye movements Figure–ground organization: the experience viewers have as to which part of an image is in front and which part of an image is in the background of a particular scene Filiform papillae: found all over the tongue; they contain somatosensory receptors rather than taste buds, so that they feel food rather than taste it Flavor: the combined sensory experience of a food, which combines its taste, its odor, its effect on the trigeminal nerve, and even visual experience Focus of expansion: the destination point in an optic flow display, from which point perceived motion derives Foliate papillae: found along the side of the tongue; they respond to all five basic tastes Forced-choice method: a psychophysical method in which a participant is required to report when or where a stimulus occurs instead of whether it was perceived Formants: frequency bands with higher amplitudes among the harmonics of a vowel sound; each individual vowel sound has a specific pattern of formants Fourier analysis: a mathematical procedure for taking any complex waveform and determining the simpler waveforms that make up that complex pattern; the simpler waves used are sine waves Fovea: an area on the retina that is dense in cones but lacks rods; when we look directly at an object, its image falls on the fovea Frequency: the number of waves per unit of time; frequency is the inverse of wavelength Frequency (sound stimulus): the number of cycles that occur in a second Functional magnetic resonance imaging (fMRI): a neuroimaging technique that generates an image of the brain on the basis of the blood levels in different areas of the brain, which correlate with activity levels in those regions Fundamental frequency: the lowest frequency in a complex sound, which determines the perceived pitch of that sound Fungiform papillae: located mostly along the edges and top of the tongue; they respond to all five basic tastes Fusiform face area: an area in the inferotemporal area of the temporal lobe that specializes in recognizing familiar faces Gate control theory: a model that allows for top-down control of the pain signal coming up the spinal cord Gelb effect: a phenomenon whereby an intensely lit black object appears to be gray or white in a homogeneously dark space General-mechanism theories: theories of speech perception that claim that the mechanisms for speech perception are the same as the mechanisms for auditory perception in general Geons: the basic units of objects, consisting of simple shapes such as cylinders and pyramids Gestalt psychology: a school of thought claiming that we view the world in terms of general patterns and well-organized structures rather than separable individual elements Glomeruli: spherical structures within the olfactory bulb where the olfactory tract forms synapses with mitral cells and tufted cells Golgi tendon organs: receptors in the tendons that measure the force of a muscle’s contraction Gradient of flow: the difference in the perception of the speeds of objects moving past us in an optic flow display Grouping: the process by which elements in a figure are brought together into a common unit or object Gustation: the sense of taste Habituation: the learning process in which animals stop responding to a repeated stimulus Hair cells: cells that have stereocilia for transducing the movement of the basilar membrane into a neural signal Haptic perception: the active use of touch to identify objects Harmonic coherence: when frequencies present in the environment resemble the possible pattern of a fundamental frequency and higher harmonics Harmonics: higher frequencies present in a complex sound that are integer multiples of the fundamental frequency (main frequency) Harmony: the pleasant sound that results when two or more musical notes are played together Sensation and Perception466 Hearing aids: electronic devices that amplify sound so that people with hearing deficits can hear sounds that otherwise would be below their thresholds Hemifield neglect (unilateral visual neglect): a condition in which a person fails to attend to stimuli on one side of the visual world (usually the left) as a consequence of neurological damage to the posterior parietal lobe Hertz (Hz): a unit of measure indicating the number of cycles per second Heterochromatic light: white light, consisting of many wavelengths Heterochromia: a condition in which a person has irises of two different colors Hit: in signal detection analysis, a hit occurs when a signal is detected when the signal is present Homunculus: a drawing of a human in which the proportions of the body parts match the relative sizes each body part has on the somatotopic map Horopter: the region in space where the two images from an object fall on corresponding locations on the two retinae Hue: the color quality of light, corresponding to the color names we use, such as orange, green, indigo, and cyan; hue is the quality of color Hue cancellation: an experiment in which observers cancel out the perception of a particular color by adding light of the opponent color Hypercolumn: a 1-mm block of V1 containing both the ocular dominance and orientation columns for a particular region in visual space Hyperopia: a condition causing an inability to focus on near objects, also called farsightedness; occurs because accommodation cannot make the lens thick enough Hyperpolarization: a change in the voltage of a neuron whereby the inside of the cell becomes more negative than it is in its resting state Illusory contours: perceptual edges that exist because of edge completion but are not actually physically present Inattentional blindness: a phenomenon in which people fail to perceive an object or event that is visible but not attended to Incus: an ossicle in the middle ear; receives vibrations from the malleus and transmits them to the stapes Induced motion: an illusion whereby one moving object may cause another object to look like it is moving Inferior colliculus: a structure in the midbrain that receives input from the superior olive Inferotemporal area: the area of the temporal lobe involved in object perception; it receives input from V4 and other areas in the occipital lobe Inferotemporal cortex: the region in the temporal lobe that receives input from the ventral visual pathway; one of its functions is object identification Information-processing approach: the view that perceptual and cognitive systems can be viewed as the flow of information from one process to another Inner hair cells: cells that are responsible for transducing the neural signal Intensity: when referring to waves, the height of a wave Interaural level difference: the difference in loudness and frequency distribution between the two ears Interaural time difference: the time interval between when a sound enters one ear and when it enters the other ear Interblobs: groups of neurons that are sensitive to orientation in vision International Phonetic Alphabet: an alphabetic convention that provides a unique symbol for each and every phoneme in use in human languages Internuclear ophthalmoplegia: a conjugate palsy resulting from damage to the brain stem region known as the medial longitudinal fasciculus Ipsilateral: literally, same (ipsi) side (lateral), meaning, in this context, that sensory information is on the same side of the nervous system as it entered Ipsilateral organization: same-side organization; in the visual system, the temporal retina projects to the same side of the brain Iris: the colored part of the eye; a muscle that controls the amount of light entering through the pupil Joint receptors: receptors found in each joint that sense information about the angle of the joint Key: the tonic note (e.g., C in a C major or minor scale) that gives a subjective sense of arrival and rest in a musical piece Koniocellular layers: layers of the lateral geniculate nucleus with very small cells that receive input from K ganglion cells (bistratified retinal ganglion cells) Koniocellular pathway (K pathway): a pathway that starts with bistratified retinal ganglion cells and projects to the koniocellular layers of the lateral geniculate nucleus Larynx (voice box): a structure that affects the pitches of sounds being made by the human vocal tract Lateral geniculate nucleus: a bilateral structure (one is present in each hemisphere) in the thalamus that relays information from the optic nerve to the visual cortex Lateral inhibition: the reduction of a response of the eye to light stimulating one receptor by stimulation of nearby receptors, caused by inhibitory signals in horizontal cells Lateral intraparietal (LIP) area: an area of the primate parietal cortex involved in the control of eye movements Law of common fate: the gestalt grouping law that states that elements that are moving together tend to be perceived as a unified group Law of good continuation: the gestalt grouping law stating that edges that are smooth are more likely to be seen as Glossary 467 continuous than edges that have abrupt or sharp angles Law of proximity: the gestalt grouping law stating that elements that are close together tend to be perceived as a unified group Law of similarity: the gestalt grouping law stating that elements that are similar to one another tend to be perceived as a unified group Law of symmetry: the gestalt grouping law that states that elements that are symmetrical to each other tend to be perceived as a unified group L-cone: the cone with its peak sensitivity to long-wavelength light, around 565 nm (yellow) Lens: the adjustable focusing element of the eye, located right behind the iris; also called the crystalline lens Light adaptation: the process whereby the visual system’s sensitivity is reduced so that it can operate in higher light levels Lightness: the psychological experience of the amount of light that gets reflected by a surface Lightness constancy: the ability to perceive the relative reflectance of objects despite changes in illumination Linear perspective: the pictorial depth cue that arises from the fact that parallel lines appear to converge as they recede into the distance Loudness: the perceptual experience of the amplitude or intensity of a sound stimulus Macrosmatic: species that are heavily dependent on their olfactory system Macula (Chapter 3): the center of the retina; the macula includes the fovea but is larger than it Macula (Chapter 14): the structure in the otolith organs that contain the receptors Macular degeneration: a disease that destroys the fovea and the area around it Magnetoencephalography (MEG): using a magnetic sensor to detect the small magnetic fields produced by electrical activity in the brain Magnitude estimation: a psychophysical method in which participants judge and assign numerical estimates to the perceived strength of a stimulus Magnocellular layers: layers of the lateral geniculate nucleus with large cells that receive input from M ganglion cells (parasol retinal ganglion cells) Magnocellular pathway (M pathway): a pathway that starts with the parasol retinal ganglion cells and projects to the magnocellular layers of the lateral geniculate nucleus Malleus: the first ossicle in the middle ear; it receives vibrations from the tympanic membrane and transmits them to the incus Manner of articulation: how airflow restriction in the vocal tract occurs Masking: refers to the difficulty in seeing one stimulus when it is quickly replaced by a second stimulus that occupies the same or adjacent spatial locations McGurk effect: a phenomenon in which vision influences the sounds people report hearing M-cone: the cone with its peak sensitivity to medium-wavelength light, around 535 nm (green) Mechanoreceptors: the sensory receptors in the skin that transduce physical movement on the skin into neural signals, which are sent to the brain Medial geniculate nucleus: a structure in the thalamus that receives auditory input from the inferior colliculus and sends output to the auditory cortex Medial intraparietal (MIP) area: an area of the posterior parietal lobe involved in the planning and control of reaching movements of the arms Meissner corpuscles: specialized transduction cells in FAI mechanoreceptors Melody: a rhythmically organized sequence of notes, which we perceive as a single musical unit or idea Merkel cells: specialized transduction cells in SAI mechanoreceptors Metamer: a psychophysical color match between two patches of light that have different sets of wavelengths Meter: the temporal pattern of sound across time Method of adjustment: a method whereby the observer controls the level of the stimulus and “adjusts” it to be at the perceptual threshold Method of constant stimuli: a method whereby the threshold is determined by presenting the observer with a set of stimuli, some above the threshold and some below it, in a random order Method of limits: stimuli are presented in a graduated scale, and participants must judge the stimuli along a certain property that goes up or down Microelectrode: a device so small that it can penetrate a single neuron in the mammalian central nervous system without destroying the cell Microsmatic: species that are less dependent on their olfactory system Middle canal (cochlear duct): one of the three chambers in the cochlea; separated from the tympanic canal by the basilar membrane; contains the organ of Corti Midget retinal ganglion cells (P cells): retinal ganglion cells that project to the parvocellular layer of the lateral geniculate nucleus; they represent 80% of ganglion cells, possess low sensitivity to light, and are sensitive to wavelength Miss: in signal detection analysis, a miss is an error that occurs when an incoming signal is not detected Mitral cells: neurons that start in the glomeruli of the olfactory bulb and project to other areas of the brain; respond to different odorants than do tufted cells Monochromatic light: light consisting of one wavelength Monocular depth cues: depth cues that require only one eye Sensation and Perception468 Moon illusion: the illusion where the moon looks larger when it is near the horizon than it does when overhead Motion: a change in position over time Motion aftereffect: a motion-based visual illusion in which a stationary object is seen as moving in the opposite direction of real or apparent motion just observed Motion parallax: a monocular depth cue arising from the relative velocities of objects moving across the retinae of a moving person Motor theory of speech perception: the theory that the core of speech perception is that the system infers the sound from the movements of the vocal tract Movement-based cues: cues about depth that can be seen with a single eye in which the inference of distance comes from motion MT (V5): an area of the occipital lobe in the dorsal pathway, specific to motion detection and perception Müller–Lyer illusion: the illusion where a line that has two lines going away at an angle looks longer than a line of the same length but the end lines angle back across the main line Muscle spindles: receptors embedded in the muscles that sense information about muscle length and therefore muscle action Music: ordered sound made and perceived by human beings, created in meaningful patterns Myopia: a condition causing an inability to focus clearly on far objects, also called nearsightedness; occurs because accommodation cannot make the lens thin enough Nasal septum: the wall of cartilage that separates the nostrils Near point: the closest distance at which an eye can focus Neural response: the signal produced by receptor cells that can then be sent to the brain Neuroimaging: technologies that allow us to map living intact brains as they engage in ongoing tasks Neuropsychology: the study of the relation of brain damage to changes in behavior Neuroscience: the study of the structures and processes in the nervous system and brain Neurotransmitter: a chemical substance neural cells release to communicate with other neural cells Nociceptive pain: pain that develops from tissue damage that causes nociceptors in the skin to fire Nociceptors: sensory receptors in the skin that, when activated, cause us to feel pain; they are found in both the epidermis and the dermis Noncorresponding points: refers to a situation in which a point on the left retina and a point on the right retina would not coincide if the two retinae were superimposed Nontasters: people who cannot detect bitter compounds except at very high concentrations Object agnosia: an acquired deficit in identifying and recognizing objects even though vision remains intact Occipital face area: an area of the brain in the occipital lobe, associated with recognizing faces as distinct from other objects Occlusion: a visual cue that occurs when one object partially hides or obstructs the view of a second object; we infer that the hidden object is farther away from us than the object that obstructs it Octave: the interval between one musical note and a note with either double the frequency or half the frequency of that note Ocular dominance column: a column within V1 that is made up of neurons that receive input from only the left eye or only the right eye Odorants: molecules our olfactory system responds to when we detect them in the air Odors: the perceptual experience of odorants, which are airborne chemical stimuli Off-center receptive fields: retinal ganglion cells that decrease their firing rate (inhibition) when light is presented in the middle of the receptive field and increase their firing rate (excitation) when light is presented in the outside or surround of the receptive field Olfaction: the sense of smell Olfactory bulb: a part of the brain just behind the nose; it is the first place in the brain where olfactory information is processed Olfactory cleft: the channel at the back of the nasal cavity that funnels air up toward the olfactory epithelium Olfactory epithelium: a mucous membrane inside each nostril of the nose that contains the receptor cells for the olfactory system Olfactory nerve (first cranial nerve): the axons of the olfactory receptor neurons that leave the nose and enter the olfactory bulb Olfactory receptor neurons: receptor cells located in the olfactory epithelium that detect specific chemicals in the air and transduce them into a neural signal Olfactory tract: the pathway leading from the olfactory bulb to other regions of the brain On-center receptive fields: retinal ganglion cells that increase their firing rate (excitation) when light is presented in the middle of the receptive field and decrease their firing rate (inhibition) when light is presented in the outside or surround of the receptive field Opponent-process theory of color perception: the theory that color perception arises from three opponent mechanisms, for red–green, blue– yellow, and black–white Opsin: the protein portion of a photopigment that captures the photon of light and begins the process of transduction; the variation in opsin determines the type of visual receptor Optic chiasm: the location in the optic tract where the optic nerve from each eye splits in half, with nasal retinae crossing over and temporal retinae Glossary 469 staying on the same side of the optic tract Optic disc: the part of the retina where the optic nerve leaves the eye and heads to the brain; along the optic disc, there are no receptor cells Optic flow: a motion depth cue that involves the relative motion of objects as an observer moves forward or backward in a scene Optic tract: the optic nerve starting at the optic chiasm and continuing into the brain Optometrist: a trained professional who specializes in diagnosing visual impairments and diseases Orbitofrontal cortex: a part of the prefrontal cortex that appears to be critical in the emotional experience of odors and integrating olfaction and taste perception, among other functions Organ of Corti: a structure on the basilar membrane that houses the hair cells that transduce sound into a neural signal Orientation column: a column within V1 that is made up of neurons with similar responses to the orientation of a shape presented to those neurons Orienting attention network (dorsal attention network): a neural system, located primarily in the parietal lobe, that allows us to engage in visual search and direct our visual attention to different locations in visual space Orienting tuning curve: a graph that demonstrates the typical response of a simple cell to stimuli or different orientations Ossicles: three small bones in the middle ear Otolith organs: organs responsible for detecting acceleration of the head and identifying when the head is being held at a tilted angle Otosclerosis: an inherited bone disease in which the ossicles, particularly the stapes, may calcify and therefore be less conductive of sound Outer hair cells: cells that sharpen and amplify the responses of the inner hair cells Overt attention: when your visual attention lines up with your direction of gaze, that is, your fovea Pacinian corpuscles: specialized transduction cells in FAII mechanoreceptors Pain: the perception and unpleasant experience of actual or threatened tissue damage Panum’s area of fusion: the region of small disparity around the horopter where the two images can be fused into a single perception Papillae: small structures that contain the taste buds Parabelt: a region of the auditory cortex, in addition to the belt area, that wraps around the auditory core regions Parahippocampal place area (PPA): an area within the inferotemporal cortex that appears to have the specific function of scene recognition Parasol retinal ganglion cells (M cells): retinal ganglion cells that project to the magnocellular layer of the lateral geniculate nucleus; they represent 10% of ganglion cells and possess high sensitivity to light Parietal insular vestibular cortex: an area in the parietal lobe that receives input from the vestibular nerve and is responsible for the perception of balance and orientation Parvocellular layers: layers of the lateral geniculate nucleus with small cells that receive input from P ganglion cells (midget retinal ganglion cells) Parvocellular pathway (P pathway): a pathway characterized by the retinal ganglion cells known as midget retinal ganglion cells Passive electroreception: the ability only to detect electric fields Perception: the process of creating conscious perceptual experience from sensory input Perceptual bistability: a phenomenon in which a static visual image leads to alternating perceptions Perceptual narrowing: the developmental process whereby regularly experienced phonemes are homed in on, with simultaneous diminishing of the ability to discriminate unfamiliar phonemes Perceptual organization: the process by which multiple objects in the environment are grouped, allowing us to identify multiple objects in complex scenes Perilymph: the fluid that fills the tympanic canal and the vestibular canal Phantom limb pain: refers to experiencing pain in a limb that has been amputated Phantom limb syndrome: continued but illusory sensory reception in a missing appendage Phantosmia: hallucinatory perception of odors Pharynx: the top portion of the throat Phase: the position in one cycle of a wave; there are 360 degrees in a single cycle of a wave Phenomenology: our subjective experience of perception Phonemes: the basic units of sound in human language Phonemic restoration effect: an illusion in which participants hear sounds that are masked by white noise, but context makes the missing sound apparent Photon: a single particle of light Photopic vision: the vision associated with the cones; it is used in the daytime, has good acuity in the fovea, and has color vision Photopigment: a molecule that absorbs light and by doing so releases an electric potential by altering the voltage in the cell Pictorial cues: information about depth that can be inferred from a static picture Pinna: the structure that collects sound and funnels it into the auditory canal Sensation and Perception470 Piriform cortex: an area in the anterior region of the temporal lobe that receives input from the olfactory bulb and is involved in olfactory processing; often considered the primary olfactory cortex Pitch: the subjective experience of sound that is most closely associated with the frequency of a sound stimulus; related to the experience of whether the sound is high or low, such as the two ends of the keyboard of a piano Place code theory: the view that different locations along the basilar membrane respond to different frequencies Place of articulation: the point along the vocal tract at which the airflow is constricted Point of subjective equality (PSE): the settings of two stimuli at which the observer experiences them as identical Point-light walker display: an experiment in which small lights are attached to the body of a person or an animal, which is then filmed moving in an otherwise completely dark environment Ponzo illusion: the illusion in which two horizontal lines are drawn one above the other; both lines are on top of two inwardly angled vertical lines; the top line, where the two vertical lines are closer together, looks longer Posterior chamber: the space between the iris and the lens; it is filled with fluid known as aqueous humor Posterior piriform cortex: a structure located in the back portion of the piriform cortex that is associated with an odor’s quality, regardless of its chemical composition Presbyopia: a condition in which incoming light focuses behind the retina, leading to difficulty focusing on close-up objects; common in older adults, in whom the lens becomes less elastic Presynaptic cells: taste receptor cells that transduce salty and sour tastes Primary auditory cortex: the first area in the auditory cortex, which receives input from the medial geniculate nucleus Primary visual cortex (V1): the area of the cerebral cortex that receives input from the lateral geniculate nucleus, located in the occipital lobe and responsible for early visual processing Proprioception: the perception of the movements and position of our limbs Prosopagnosia: face agnosia, resulting in a deficit in perceiving faces Protanopia: a lack of L-cones, leading to red–green deficiency; this trait is sex- linked and thus more common in men Pruriceptors: receptors in our skin that respond to mild irritants by producing itch sensations Psychophysical scale: a scale on which people rate their psychological experiences as a function of the level of a physical stimulus Psychophysics: the study of the relation between physical stimuli and perception events Pupil: an opening in the middle of the iris Pupillary reflex: an automatic process by which the iris contracts or relaxes in response to the amount of light entering the eye; the reflex controls the size of the pupil Pure tone: a sound wave in which changes in air pressure follow a sine wave pattern Purkinje shift: the observation that short wavelengths tend to be relatively brighter than long wavelengths in scotopic vision versus photopic vision Quality: a value that changes but does not make the value larger or smaller Random-dot stereograms: stereograms in which the images consist of a randomly arranged set of black and white dots, with the left-eye and right-eye images arranged identically except that some of the dots are moved to the left or the right in one of the images, creating either a crossed or an uncrossed disparity Rapid serial visual presentation (RSVP) paradigm: a method of studying attention in which a series of stimuli appear rapidly in time at the same point in visual space Rate of approach: the measure of whether a predator is approaching a target or receding from it Real motion: motion in the world created by continual change in the position of an object relative to some frame of reference Receiver-operating characteristic (ROC) curve: in signal detection theory, a plot of false alarms versus hits for any given sensitivity, indicating all possible outcomes for a given sensitivity Receptive field: a region of adjacent receptors that will alter the firing rate of a cell that is higher up in the sensory system; the term can also apply to the region of space in the world to which a particular neuron responds Receptor cells: taste receptor cells that transduce sweet tastes, umami tastes, and bitter tastes Receptors: specialized sensory neurons that convert physical stimuli into neural responses Recognition: the ability to match a presented item with an item in memory Recognition by components: a theory stating that object recognition occurs by representing each object as a combination of basic units (geons) that make up that object; we recognize an object by the relation of its geons Reichardt detectors: neural circuits that enable the determination of direction and speed of motion by delaying input from one receptive field, to determine speed, to match the input of another receptive field, to determine direction Reissner’s membrane: the membrane that separates the vestibular and middle canals Relative height: a visual cue in which objects closer to the horizon are seen as more distant Relative size: the fact that the more distant an object is, the smaller the image will be on the retina Repetition blindness: the failure to detect the second target in an RSVP task when the second target is identical to the first Glossary 471 one; like attentional blink, it occurs when the second target is presented 500 ms or less after the first target Representation: the storage and/or reconstruction of information in memory when that information is not in use Response compression: as the strength of a stimulus increases, so does the perceptual response, but the perceptual response does not increase by as much as the stimulus increases Response expansion: as the strength of a stimulus increases, the perceptual response increases even more Retina: the paper-thin layer of cells at the back of the eye where transduction takes place Retinal: a derivative of vitamin A that is part of a photopigment Retinal image: the light projected onto the retina Retinitis pigmentosa: an inherited progressive degenerative disease of the retina that may lead to blindness Retinotopic map: a point-by-point relation between the retina and V1 Reverberation time: the difference between the onset of direct sound and the first onset of indirect sound Rhythm: the temporal patterning of music, including the tempo, the beat, and the meter Rods: photoreceptors at the periphery of the retina; they are very light sensitive and specialized for night vision Rostral core: an area in the auditory core region of the auditory cortex Rostrotemporal core: an area, in addition to the rostral core, in the auditory core region of the auditory cortex Round window: a soft tissue substance at the base of the tympanic canal whose function is as an “escape” valve for excess pressure from loud sounds that arrive in the cochlea Ruffini endings: specialized transduction cells in SAII mechanoreceptors SAI mechanoreceptors: slow-adapting receptors using Merkel cells, with small receptive fields, densely packed near the surface of the skin SAII mechanoreceptors: slow-adapting receptors using Ruffini endings, with large receptive fields, more widely distributed, deeper in the skin Saccades: the most common and rapid of eye movements; sudden eye movements that are used to look from one object to another Saturation: the purity of light Scale: a set of ordered notes starting at one note and ending at the same note one octave higher Sclera: the outside surface of the eye; a protective membrane covering the eye that gives the eye its characteristic white appearance S-cone: the cone with its peak sensitivity to short-wavelength light, around 420 nm (blue) Scotoma: an area of partially or completely destroyed cells, resulting in a blind spot in a particular region of the visual field Scotopic vision: the operation of the visual system associated with the rods; it has relatively poor acuity and no color ability but is very sensitive to light Scoville scale: a measure of our detection of the amount of an ingredient called capsaicin in chili peppers Segregation: the process of distinguishing two objects as being distinct or discrete Selective attention: the processes of attention that allow us to focus on one source when many are present Semicircular canals: three tubes located in the inner ear responsible for the signaling of head rotation Semitones: the 12 equivalent intervals or notes within each octave Sensation: the registration of physical stimuli on sensory receptors Sensitivity: the ability to perceive a particular stimulus; it is inversely related to threshold Sensitivity (signal detection theory): the ease or difficulty with which an observer can distinguish signal from noise Sensorineural hearing loss: permanent hearing loss caused by damage to the cochlea or auditory nerve or the primary auditory cortex Shadows: a depth cue arising because an object is in front of its shadow; the angle of the shadow can provide information about how far the object is in front of the background Signal detection theory: the theory that in every sensory detection or discrimination, there is both sensory sensitivity to the stimulus and a criterion used to make a cognitive decision Simple cells: V1 neurons that respond to stimuli with particular orientations to objects within their receptive field; the preferred orientation of a simple cell is the stimulus orientation that produces the strongest response Simultagnosia: a deficit in perceiving more than one object at a time Simultaneous color contrast: a phenomenon that occurs when our perception of one color is affected by a color that surrounds it Size constancy: the perception of an object as having a fixed size, despite the change in the size of the visual angle that accompanies changes in distance Size–arrival effect: bigger approaching objects are seen as being more likely to collide with the viewer than smaller approaching objects Size–distance invariance: the relation between perceived size and perceived distance, whereby the perceived size of an object depends on its perceived distance, and the perceived distance of an object may depend on its perceived size Smooth-pursuit eye movements: voluntary tracking eye movements Somatosensory cortex: an area in the parietal lobe of the cerebral cortex devoted to processing the information coming from the skin senses Somatotopic map: a feature whereby the skin of the body maps onto the surface of the primary somatosensory cortex in a systematic way Sensation and Perception472 Sound stimulus: the periodic variations in air pressure traveling out from the source of the variations Sound waves: the waves of pressure changes that occur in the air as a function of the vibration of a source Spatial segregation: the process whereby sounds that are coming from the same location are grouped together, whereas sounds that are coming from different locations are not grouped together Special-mechanism theories: theories of speech perception that claim that the mechanisms for speech perception are distinct and unique relative to the mechanisms for auditory perception in general Spectral reflectance: the ratio of light reflected by an object at each wavelength Spectral segregation: the process whereby sounds that overlap in harmonic structure are grouped together, whereas sounds that do not overlap in harmonic structure are not grouped together Spectral shape cue: the change in a sound’s frequency envelope created by the pinnae; it can be used to provide information about the elevation of a sound source Spinothalamic pathway: a pathway for the nociceptors (pain) and thermoreceptors (temperature) that travels up the contralateral side of the spinal column; it does not synapse in the brain until the ventral posterior nucleus of the thalamus Stapedius: the muscle that is attached to the stapes Stapes: an ossicle in the middle ear; it receives vibrations from the incus and transmits them to the oval window of the inner ear Stereocilia: the hairlike parts of the hair cells on the top of the inner and outer hair cells Stereopsis: the sense of depth we perceive from the visual system’s processing of the comparison of the two different images from each retina Stevens’s power law: a mathematical formula that describes the relationship between stimulus intensity and our perception; it allows for both response compression and response expansion Stimulus: an element of the world around us that impinges on our sensory systems Stimulus onset asynchrony: the difference in time between the occurrence of one stimulus and the occurrence of another, in this case, the cue and the target Stimulus salience: refers to the features of objects in the environment that attract our attention Substantia gelatinosa: the region in the dorsal horn where neurons meet Subtractive color mixing: color mixing in which a new color is made by the removal of wavelengths from a light with a broad spectrum of wavelengths Superior colliculus: a structure located at the top of the brain stem, just beneath the thalamus, whose main function in mammals (including humans) is the control of eye movements Superior olive: a structure in the brain stem that receives input from the inner hair cells and from the cochlear nucleus Supertasters: people who are extremely sensitive to bitter tastes; they usually do not like foods with many bitter compounds Supporting cells: cells that provide metabolic supplies to the olfactory receptor neurons Synesthesia: a condition in which a stimulus in one modality consistently triggers a response in another modality Tactile agnosia: an inability to identify objects by touch Tapetum: a reflective layer behind the receptors of nocturnal animals that bounces light not caught by receptors back into the retina Target range: the distance of a predator from its potential target, determined by timing an echo’s return Tastants: molecules recognized by taste receptors that induce responses in taste receptors on the tongue Taste: the perception of the transduction of tastants along the surface of the tongue Taste buds: small structures located along the surface of the tongue or mouth that contain the receptor cells Taste receptor cells: cells within the taste buds that transduce tastants into a neural signal Tasters: people who can detect bitter compounds Tectorial membrane: a membrane that rests above the hair cells within the organ of Corti Tempo: the pace at which a piece of music is played Temporal code theory: the view that frequency representation occurs because of a match between sound frequency and the firing rates of the auditory nerve Temporal segregation: the process whereby sounds that are linked in time are grouped together, whereas sounds that are not correlated with one another are not grouped together Tensor tympani: the muscle that is attached to the malleus Texture gradient: a monocular depth cue that occurs because textures become finer as they recede in the distance Thermoreception: the ability to sense changes in temperature on the skin Thermoreceptors: the sensory receptors in the skin that signal information about the temperature as measured on the skin Timbre: the perceived sound differences between sounds with the same pitch but possessing different higher harmonics Time to collision: the estimate of the time it will take for an approaching object to contact another Tinnitus: a condition in which people perceive sounds even when none are present Tip-of-the-nose phenomenon: a phenomenon that occurs when a Glossary 473 person is familiar with an odor but cannot recall its name, despite feeling as if he or she can Tonotopic organization: the organization of neurons within a region in the brain according to the different frequencies to which they respond Top-down processing: a process whereby our existing knowledge of objects influences how we perceive them Topographic agnosia: a deficit in recognizing spatial landscapes, related to damage to the parahippocampal place area Trachea (windpipe): the tube bringing air to and from the mouth Transduction: the process of converting a physical stimulus into an electrochemical signal Transmagnetic stimulation (TMS): a procedure in which a magnetic coil is used to stimulate electrically a specific region of the brain Transposition: the process through which one can create multiple versions of a melody that start on different notes but contain the same intervals or sequence of changes in notes Trapezoid body: a structure in the brain stem that plays a role in determining the direction of sounds Trichromatic theory of color vision: the theory that the color of any light is determined by the output of the three cone systems in our retinae Trigeminal nerve: a nerve that is associated with the feel of odorants; also known as the fifth cranial nerve Tritanopia: a lack of S-cones, leading to blue–yellow color deficiency; this trait is rare and is not sex-linked Tuberous receptor: the organ that contains the hair cells that detect electric fields, used in active electroreception Tufted cells: neurons that start in the glomeruli of the olfactory bulb and project to other areas of the brain; they respond to different odorants than do mitral cells Turbinates: bony knots of tissue that serve to disperse air within the nasal cavity Two-point touch threshold: the minimum distance at which two touches are perceived as two touches and not one Tympanic canal: one of the three chambers in the cochlea; separated from the middle canal by the basilar membrane Tympanic membrane: a thin elastic sheet that vibrates in response to sounds coming through the external auditory canal; commonly known as the eardrum Unconscious inference: perception is not adequately determined by sensory information, so an inference or educated guess is part of the process; this inference is not the result of active problem solving but rather a nonconscious cognitive process Uncrossed disparity: the direction of disparity for objects that are farther from the viewer than the horopter (the image of the object in the left eye is to the left of the position of the image of the object in the right eye) Unilateral dichromacy: the presence of dichromacy in one eye but normal trichromatic vision in the other Unique colors: colors that can be described only with a single color term—red, green, blue, and yellow Univariance: the principle whereby any single cone system is colorblind, in the sense that different combinations of wavelength and intensity can result in the same response from the cone system Unvoiced consonant: a consonant that is produced without using the vocal chords Uvula: a flap of tissue at the top of the throat that can close off the nasal cavity V2: the second area in the visual cortex that receives input; often considered the area that starts with visual associations rather than processing the input (sometimes called the prestriate cortex) V4: an area of the occipital lobe involved in both color vision and shape perception Ventral: in or toward the front of the body; in the head, it means at the bottom or toward the bottom Ventral pathway: starts with midget and bistratified retinal ganglion cells and continues through the visual cortex into the inferotemporal cortex in the temporal lobe; often called the “what” pathway, as it codes for object identification as well as color vision Ventral posterior nucleus of the thalamus: an area in the thalamus that receives input from both the dorsal column–medial lemniscal pathway and the spinothalamic pathway Ventral root: the end of the spinal cord where motor information leaves the spinal cord Vergence: the inward bending of the eyes when looking at closer objects Vestibular canal: one of the three chambers in the cochlea; separated from the middle canal by Reissner’s membrane Vestibular complex: a brain stem area that receives input from the vestibular nerve and sends the information to the forebrain Vestibular system: the sensory system responsible for the perception of balance and acceleration, housed in the semicircular canals and otolith organs, both located adjacent to the inner ear Viewpoint invariance: the perception that an object does not change when an observer sees the object from a new vantage point Virtual reality: a computer-generated photograph, image, or environment that can be interacted with in an apparently real way Visual angle: the angle of an object relative to the observer’s eye Visual capture: circumstances where visual input can dominate the input from other senses when they conflict Visual search: looking for a specific target among distracting objects Visual spectrum (visible spectrum): the band of wavelengths from 400 Sensation and Perception474 to 700 nm that people with the most common classes of cones vision can detect Voice area: an area located in the superior temporal sulcus that responds to the sound of the human voice, but less so to other stimuli Voiced consonant: a consonant that is produced using the vocal chords Voicing: whether the vocal cords are vibrating or not Voicing-onset time: the production of certain consonants (called stop consonants) in which there is a difference between the first sound of the phoneme and the movement of the vocal cords (called voicing) Vowels: speech sounds made with unrestricted airflow Warm fibers: thermoreceptors that fire in response to warmer temperatures (above 36 °C) as measured on the skin Wavelength: the distance between two adjacent peaks in a repeating wave; different forms of electromagnetic energy are classified by their wavelengths Weber’s law: a just-noticeable difference between two stimuli is related to the magnitude or strength of the stimuli Wernicke’s aphasia: a form of aphasia resulting from damage to Wernicke’s area, causing a deficit in language comprehension Wernicke’s area: an important area in speech comprehension, located in the left temporal lobe Word segmentation: the ability of speakers of a language to correctly perceive boundaries between words Zero disparity: the situation in which retinal images fall along corresponding points, which means that the object is along the horopter Zonule fibers: fibers that connect the lens to the choroid membrane References 475 REFERENCES Abdelaziz, A. H., & Kolossa, D. (2016). General hybrid framework for uncertainty-decoding-based automatic speech recognition systems. Speech Communication, 79, 1–13. Adelson, E. H. (1993). Perceptual organization and the judgment of brightness. Science, 262, 2042–2044. Aladdin, Y., Scozzafava, J., Muayqil, T., & Saqqur, M. (2008). Pearls & oysters: Oculopalatal tremor with one-and-a- half syndrome after pontine hemorrhage. Neurology, 71, 39–41. Albright, T. D. (1984). Direction and orientation selectivity of neurons in visual area MT of the macaque. Journal of Neurophysiology, 52, 1106–1130. Alkaladi, A., & Zeil, J. (2014). Functional anatomy of the fiddler crab compound eye (Uca vomeris: Ocypodidae, Brachyura, Decapoda). Journal of Comparative Neurology, 522, 1264–1283. Alpern, M., Kitahara, K., & Krantz, D. H. (1983). Perception of color in unilateral tritanopia. Journal of Physiology, 335, 683–697. Alpern, M., Lee, G. B., Maaseidvaag, F., & Miller, S. S. (1971). Colour vision in blue-cone “monochromacy.” Journal of Physiology, 212, 211–233. Alvarado, J. C., Vaughan, J. W., Stanford, T. R., & Stein, B. E. (2007). Multisensory versus unisensory integration: Contrasting modes in the superior colliculus. Journal of Neurophysiology, 97, 3193–3205. Ambady, N., & Bharucha, J. J. (2009). Culture and the brain. Current Directions in Psychological Science, 18, 342–345. Anderson, B. A., Laurent, P. A., & Yantis, S. (2013). Reward predictions bias attentional selection. Frontiers in Human Neuroscience, 7, 262. Anderson, B. A., & Yantis, S. (2013). Persistence of value- driven attentional capture. Journal of Experimental Psychology: Human Perception and Performance, 39, 6–9. Anderson, S., White-Schwoch, T., Parbery-Clark, A., & Kraus, N. (2013). A dynamic auditory-cognitive system supports speech-in-noise perception in older adults. Hearing Research, 300, 18–32. Anstis, S. (2012). The furrow illusion: Peripheral motion becomes aligned with stationary contours. Journal of Vision, 12, 1–11. Anstis, S., & Rogers, B. (2011). Illusory rotation of a spoked wheel. i-Perception, 2, 720–723. Arroyo, J. G. (2006). A 76-year-old man with macular degeneration. JAMA, 295, 2394–2406. Atkinson, J., Braddick, O., & Moar, K. (1977). Development of contrast sensitivity over the first 3 months of life in the human infant. Vision Research, 17(9), 1037–1044. Attems, J., Walker, L., & Jellinger, K. A. (2014). Olfactory bulb involvement in neurodegenerative diseases. Acta Neuropathologica, 127, 459–475. Babadi, B., Casti, A., Xiao, Y., Kaplan, E., & Paninski, L. (2010). A generalized linear model of the impact of direct and indirect inputs to the lateral geniculate nucleus. Journal of Vision, 10, 22. Balkwill, L.-L., & Thompson, W. F. (1999). A cross-cultural investigation of the perception of emotion in music: Psychophysical and cultural cues. Music Perception, 17, 43–64. Banks, M. S., & Bennett, P. J. (1988). Optical and photoreceptor immaturities limit the spatial and chromatic vision of human neonates. JOSA A, 5(12), 2059–2079. Bannister, J. (2016). Chili peppers 101. Retrieved from https://cpi.nmsu.edu/chile-pepper-news-and-info-center Barlow, H. B. (1953). Summation and inhibition in the frog’s retina. Journal of Physiology, 119, 69–88. Barlow, H. B. (1995). The neuron in perception. In M. S. Gazzaniga (Ed.), The cognitive neurosciences (pp. 415–434). Cambridge, MA: MIT Press. Barlow, H. B., Blakemore, C., & Pettigrew, J. D. (1967). The neural mechanism of binocular depth discrimination. Journal of Physiology, 193, 327–342. Barry, M. P., & Dagnelie, G. (2016). Hand-camera coordination varies over time in users of the Argus® II retinal prosthesis system. Frontiers in Systems Neuroscience, 10, 41. Bartoshuk, L. M. (1971). The chemical senses: I. Taste. In J. W. Kling & L. A. Riggs (Eds.), Experimental psychology (3rd ed.). New York, NY: Holt, Rinehart & Winston. Bartoshuk, L. M., Duffy, V. B., & Miller, I. J. (1994). PTC/ PROP tasting: Anatomy, psychophysics, and sex effects. Physiology & Behavior, 56, 1165–1171. Beck, D. M., Muggleton, N., Walsh, V., & Lavie, N. (2006). Right parietal cortex plays a critical role in change blindness. Cerebral Cortex, 16, 712–717. Beckers, G., & Zeki, S. (1995). The consequences of inactivating areas V1 and V5 on visual motion perception. Brain, 118, 49–60. Belin, P., Bestelmeyer, P. E. G., Latinus, M., & Watson, R. (2011). Understanding voice perception. British Journal of Psychology, 102, 711–725. Bellugi, U., Klima, E. S., & Hickok, G. (2010). Brain organization: Clues from deaf signers with left or right hemisphere lesions. In L. Clara (Ed.), Gesture and word (pp. 2–17). Lisbon, Portugal: Editorial Caminho. Benedetti, F. (1985). Processing of tactile spatial information with crossed fingers. Journal of Experimental Psychology: Human Perception and Performance, 11, 517–525. Sensation and Perception476 Bennett, C. M., Wolford, G. L., & Miller, M. B. (2009). The principled control of false positives in neuroimaging. Social, Cognitive, and Affective Neuroscience, 4, 417–422. Beran, M. J. (2012). We cannot know what it is like to be a bat, but we can know better what it is like to see like a bat. PsychCRITIQUES, 57, 38. Beranek, L. L. (2003). Subjective rank-orderings and acoustical measurements for fifty-eight concert halls. Acta Acustica Acustica, 89, 494–508. Bhalla, M., & Proffitt, D. R. (1999). Visual-motor recalibration in geographical slant perception. Journal of Experimental Psychology: Human Perception and Performance, 25, 1076–1096. Biederman, I. (1987). Recognition-by-components: A theory of human image understanding. Psychological Review, 94, 115–147. Birch, E. E. (1993). Introduction to binocular vision. In K. Simons (Ed.), Early visual development: Normal and abnormal. New York, NY: Oxford University Press. Birch, J. (2001). Diagnosis of defective colour vision (2nd ed.). New York, NY: Elsevier. Bisiach, E., & Luzzatti, C. (1978). Unilateral neglect of representational space. Cortex, 14, 129–133. Blackwell, H. R. (1953). Psychophysical thresholds: Experimental studies of methods of measurement (Bulletin No. 36). Ann Arbor: University of Michigan Engineering Research Institute. Blake, R., & Shiffrar, M. (2007). Perception of human motion. Annual Review of Psychology, 58, 47–73. Blake, R., Cool, S. J., & Crawford, M. L. J. (1974). Visual resolution in the cat. Vision Research, 14(11), 1211–1217. Bleyenheuft, Y., & Thonnard, J. L. (2009). Development of touch. Scholarpedia, 4(11), 7958. Bohn, K. M., Smarsh, G. C., & Smotherman, M. S. (2013). Social context evokes rapid changes in bat song syntax. Animal Behavior, 85, 1485–1491. Born, R. T., & Bradley, D. C. (2005). Structure and function of visual area MT. Annual Review of Neuroscience, 28, 157–189. Bornstein, M. H., Kessen, W., & Weiskopf, S. (1976). Color vision and hue categorization in young human infants. Journal of Experimental Psychology: Human Perception and Performance, 2(1), 115–129. Bouchard, S., Côté, S., St-Jacques, J., Robillard, G., & Renaud, P. (2006). Effectiveness of virtual reality exposure in the treatment of arachnophobia using 3D games. Technology and Health Care, 14(1), 19–27. Bowmaker, J. K. (1998). Visual pigments and molecular genetics of color blindness. News in Physiological Science, 13, 63–69. Bregman, A. S. (1990). Auditory scene analysis: The perceptual organization of sound. Cambridge, MA: Bradford. Bregman, A. S. (2005). Auditory scene analysis and the role of phenomenology in experimental psychology. Canadian Psychology, 46, 32–40. Breitmeyer, B. G., & Hanif, W. (2008). “Change of mind” within and between nonconscious (masked) and conscious (unmasked) visual processing. Consciousness and Cognition, 17, 254–266. Brendel, E., DeLucia, P. R., Hecht, H., Stacy, R. L., & Larsen, J. T. (2012). Threatening pictures induce shortened time-to-contact estimates. Attention, Perception, & Psychophysics, 74, 979–987. Bressler, N. M. (2004). Age-related macular degeneration is the leading cause of blindness. JAMA, 291, 1900–1901. Brett, J. A. (2016). Thinking local about self-driving cars: A local framework for autonomous vehicle development in the United States. Master’s thesis, University of Washington. Retrieved from https:// digital.lib.washington.edu/researchworks/bitstream/ handle/1773/36852/Brett_washington_0250O_16021 ?sequence=1&isAllowed=y Bridgeman, B., & Stark, L. (1991). Ocular proprioception and efference copy in registering visual direction. Vision Research, 31, 1903–1913. Brincat, S., & Connor, C. E. (2004). Underlying principles of visual shape selectivity in posterior inferotemporal cortex. Nature Neuroscience, 7, 880–886. Brown, A. M., & Teller, D. Y. (1989). Chromatic opponency in 3-month-old human infants. Vision Research, 29(1), 37–45. Bruce, C., Desimone, R., & Gross, C. G. (1981). Visual properties of neurons in a polysensory area in superior temporal sulcus of the macaque. Journal of Neurophysiology, 46, 369–384. Buck, L. B., & Axel, R. (1991). A novel multigene family may encode odorant receptors: A molecular basis for odor recognition. Cell, 65, 175–187. Buetti, S., Tamietto, M., Hervais-Adelman, A., Kerzel, D., de Gelder, B., & Pegna, A. J. (2013). Dissociation between goal-directed and discrete response localization in a patient with bilateral cortical blindness. Journal of Cognitive Neuroscience, 25, 1769–1775. Busch, N. A. (2013). The fate of object memory traces under change detection and change blindness. Brain Research, 1520, 107–115. Busigny, T., & Rossion, B. (2011). Holistic processing impairment can be restricted to faces in acquired prosopagnosia: Evidence from the global/local Navon effect. Journal of Neuropsychology, 5, 1–14. Buzás, P., Kóbor, P., Petykó, Z., Telkes, I., Martin, P. R., & Lénárd, L. (2013). Receptive field properties of color opponent neurons in the cat lateral geniculate nucleus. Journal of Neuroscience, 33, 1451–1461. Cain, W. S. (1979). To know with the nose: Keys to odor identification. Science, 203, 467–470. References 477 Cain Miller, C., & Wald, M. L. (2013, May 30). Self-driving cars for testing are supported by the U.S. The New York Times. Retrieved from http://www.nytimes .com/2013/05/31/technology/self-driving-cars-for-testing- are-supported-by-us.html?pagewanted=all&_r=0 Caminiti, F., Ciurleo, R., Bramanti, P., & Marino, S. (2013). Persistent anosmia in a traumatic brain injury patient: Role of orbitofrontal cortex. Brain Injury, 27, 1715–1718. Carroll, J., Neitz, M., Hofer, H., Neitz, J., & Williams, D. R. (2004). Functional photoreceptor loss revealed with adaptive optics: An alternate cause of color blindness. Proceedings of the National Academy of Sciences of the United States of America, 101, 8461–8466. Cartwright, M. (2013). Ancient Greek music. In Ancient history encyclopedia. Retrieved from https://www .ancient.eu/Greek_Music Centers for Disease Control and Prevention. (2017). Hearing loss in children: Data and statistics. Retrieved from https://www.cdc.gov/ncbddd/hearingloss/data.html Chabris, C., & Simons, D. (2009). The invisible gorilla. New York, NY: Broadway. Chan, B. L., Witt, R., Charrow, A. P., Magee, A., Howard, R., Pasquina, P. F., Heilman, K. M., & Tsao, J. W. (2007). Mirror therapy for phantom limb pain. New England Journal of Medicine, 357(21), 2206–2207. Chandrashekar, J., Hoon, M. A., Ryba, N. J. P., & Zuker, C. S. (2006). The receptors and cells for mammalian taste. Nature, 444, 288–294. Chasin, M. (2011, May). Setting hearing aids differently for different languages. Seminars in Hearing, 32(2), 182–188. Chasin, M. (2012). Sentence final hearing aid gain requirements of some non-English languages. Ajustements spécifiques de gain des appareils auditifs pour les finales de phrases de certaines langues autres que l’anglais. Canadian Journal of Speech-Language Pathology and Audiology, 36(3), 196–203. Chasin, M., & Hockley, N. S. (2014). Some characteristics of amplified music through hearing aids. Hearing Research, 308, 2–12. Chen, M.-C., Tsai, P.-L., Huang, Y.-T., & Lin, K. (2013). Pleasant music improves visual attention in patients with unilateral neglect after stroke. Brain Injury, 27, 75–82. Cherry, E. C. (1953). Some experiments on the recognition of speech, with one and with two ears. Journal of the Acoustical Society of America, 25, 975–979. Chino, Y. M., Smith, E. L., III, Hatta, S., & Cheng, H. (1997). Postnatal development of binocular disparity sensitivity in neurons of the primate visual cortex. Journal of Neuroscience, 17, 296–307. Chiu, Y.-C., & Yantis, S. (2009). A domain-independent source of cognitive control for task sets: Shifting spatial attention and switching categorization rules. Journal of Neuroscience, 29, 3930–3938. Chow, A., Chow, V., Packo, K., Pollack, J., Peyman, G., & Schuchard, R. (2004). The artificial silicone retina microchip for the treatment of vision loss from retinitis pigmentosa. Archives of Ophthalmology, 122, 1156–1157. Chrastil, E. R. (2013). Neural evidence supports a novel framework for spatial navigation. Psychonomic Bulletin & Review, 20, 208–227. Chun, M. M., Golomb, J. D., & Turke-Browne, N. B. (2011). A taxonomy of external and internal attention. Annual Review of Psychology, 62, 73–101. Coben, J. H., & Zhu, M. (2013). Keeping an eye on distracted driving. JAMA, 309(9), 877–878. Colby, C. L., Duhamel, J.-R., & Goldberg, M. E. (1996). Visual, pre-saccadic, and cognitive activation of single neurons in monkey lateral intraparietal area. Journal of Neurophysiology, 76, 2841–2851. Cole, B. L. (2016). Colour vision and driving. Clinical Experimental Ophthalmology, 99, 484–487. Collignon, O., Dormal, G., Albouy, G., Vandewalle, G., Voss, P., Phillips, C., & Lepore, F. (2013). Impact of blindness onset on the functional organization and the connectivity of the occipital cortex. Brain, 136, 2769–2783. Colombo, J. (2002). Infant attention grows up: The emergence of a developmental cognitive neuroscience perspective. Current Directions in Psychological Science, 11, 196–200. Compton, W. M., Jones, C. M., & Baldwin, G. T. (2016). Relationship between nonmedical prescription-opioid use and heroin use. New England Journal of Medicine, 374(2), 154–163. Conard, N. J., Malina, M., & Münzel, S. C. (2009). New flutes document the earliest musical tradition in southwestern Germany. Nature, 460, 737–740. Coren, S., & Girgus, J. S. (1978). Seeing is deceiving: The psychology of visual illusions. Hillsdale, NJ: Erlbaum. Cornsweet, T. N. (1970). Visual perception. New York, NY: Academic Press. Crystal, S. R., & Bernstein, I. L. (1995). Morning sickness: Impact on offspring salt preference. Appetite, 25, 231–240. Culham, J. C., Cavina-Pratesi, C., & Singhal, A. (2006). The role of parietal cortex in visuomotor control: What have we learned from neuroimaging? Neuropsychologia, 44, 2668–2684. Curcio, C. A., Sloan, K. R., Kalina, R. E., & Hendrickson, A. E. (1990). Human photoreceptor topography. Journal of Comparative Neurology, 292, 497–523. Curtiss, S. (2014). Genie: A psycholinguistic study of a modern-day wild child. New York, NY: Academic Press. Dai, P., & Soon, I. Y. (2013). An improved model of masking effects for robust speech recognition system. Speech Communication, 55(3), 387–396. Davis, K. D., Taylor, S. J., Crawley, A. P., Wood, M. L., & Mikulis, D. J. (1997). Functional MRI of pain- and Sensation and Perception478 attention-related activations in the human cingulate cortex. Journal of Neurophysiology, 77, 3370–3380. Daw, N. (2009). The foundations of development and deprivation in the visual system. Journal of Physiology, 587, 2769–2773. Day, R. H. (1990). The Bourdon illusion in haptic space. Perception and Psychophysics, 47, 400–404. de Gelder, B. (2010). Uncanny sight in the blind. Scientific American, 302, 60–65. de Gelder, B., Tamietto, M., van Boxtel, G., Goebel, R., Sahraie, A., Van den Stock, J., . . . Pegna, A. (2008). Intact navigation skills after bilateral loss of striate cortex. Current Biology, 18, R1128–R1129. De Valois, R. L. (1960). Color vision mechanisms in the monkey. Journal of General Physiology, 43, 115–128. De Valois, R. L. (1965). Behavioral and electrophysiological studies of primate vision. In W. D. Ness (Ed.), Contributions to sensory physiology (Vol. 1). New York, NY: Academic Press. De Valois, R. L. (2004). Neural coding of color. In L. M. Chalupa & J. S. Werner (Eds.), The visual neurosciences (Vol. 2, pp. 1003–1016). Cambridge, MA: MIT Press. De Valois, R. L., Abramov, I., & Jacobs, G. H. (1966). Analysis of response patterns of LGN cells. Journal of the Optical Society of America, 56, 966–977. de Wijk, R. A., & Cain, W. S. (1994). Odor quality: Discrimination versus free and cued identification. Perception and Psychophysics, 56, 12–18. DeCasper, A. J., & Fifer, W. P. (1980). Of human bonding: Newborns prefer their mothers’ voices. Science, 208, 1174–1176. Delk, J. L., & Fillenbaum, S. (1965). Differences in perceived color as a function of characteristic color. American Journal of Psychology, 78, 290–293. Delogu, F., Huddas, C., Steven, K., Hachem, S., Lodhia, L., Fernandez, R., & Logerstedt, M. (2016). A dissociation between recognition and hedonic value in caloric and non-caloric carbonated soft drinks. Frontiers in Psychology, 7, 36. DeLucia, P. R. (2013). Effects of size on collision perception and implications for perceptual theory and transportation safety. Current Directions in Psychological Science, 22, 199–204. DeLucia, P. R., Meza-Arroyo, M., Baures, R., Ranjit, M., Hsiang, S., & Gorman, J. C. (2016). Continuous response monitoring of relative time-to-contact judgments: Does effective information change during an approach event? Ecological Psychology, 28, 1–22. Delwiche, J. F., Buletic, Z., & Breslin, P. A. S. (2001). Covariation in individuals’ sensitivities to bitter compounds: Evidence supporting multiple receptor/transduction mechanisms. Perception and Psychophysics, 63, 761–776. DeMarco, P. J., Jr., Yarbrough, G. L., Yee, C. W., McLean, G. Y., Sagdullaev, B. T., Ball, S. L., & McCall, M. A. (2007). Stimulation via a subretinally placed prosthetic elicits central activity and induces a trophic effect on visual responses. Investigative Ophthalmology & Visual Science, 48, 916–926. Dennett, D. C. (1991). Consciousness explained. Boston, MA: Little, Brown. Deutsch, D. (1974). An auditory illusion. Nature, 251, 307–309. Deutsch, D. (1975a). Musical illusions. Scientific American, 233, 92–104. Deutsch, D. (1975b). Two-channel listening to musical scales. Journal of the Acoustical Society of America, 57, 1156–1160. Deutsch, D. (1986). A musical paradox. Music Perception, 3, 275–280. Deutsch, D. (2013). The psychology of music (3rd ed.). San Diego, CA: Elsevier. Devos, M., Patte, F., Rouault, J., Laffort, P., & Van Gemert, L. J. (1990). Standardized human olfactory thresholds. Oxford, UK: Oxford University Press. Dewald, A. D., Sinnett, S., & Doumas, L. A. A. (2013). A window of perception when diverting attention? Enhancing recognition for explicitly presented, unattended, and irrelevant stimuli by target alignment. Journal of Experimental Psychology: Human Perception and Performance, 39(5), 1304–1312. DeWitt, L. A., & Samuel, A. G. (1990). The role of knowledge-based expectation in music perception: Evidence for musical restoration. Journal of Experimental Psychology: General, 119, 123–144. DeYoe, E. A., & Van Essen, D. C. (1988). Concurrent processing streams in monkey visual cortex. Trends in Neurosciences, 11, 219–226. Diehl, R. L., Lotto, A. J., & Holt, L. L. (2004). Speech perception. Annual Review of Psychology, 55, 149–179. Djordjevic, J., Lundstrom, J. N., Clement, F., Boyle, J. A., Pouliot, S., & Jones-Gotman, M. (2008). A rose by any other name: Would it smell as sweet? Journal of Neurophysiology, 99, 386–393. Djordjevic, J., Zatorre, R. J., Petrides, M., Boyle, J., & Jones-Gotman, M. (2005). Functional neuroimaging of odor imagery. NeuroImage, 24, 791–801. Doerschner, K., Kersten, D., & Schrater, P. (2009). Rapid classification of surface reflectance from image velocities. Computer Analysis of Images and Patterns, Lecture Notes in Computer Science, 5702, 856–864. Domínguez-Borràs, J., Armony, J. L., Maravita, A., Driver, J., & Vuilleumier, P. (2013). Partial recovery of visual extinction by Pavlovian conditioning in a patient with hemispatial neglect. Cortex, 49, 891–898. Downing, P. E., Jiang, Y., Shuman, M., & Kanwisher, N. (2001). Cortical area selective for visual processing of the human body. Science, 293, 2470–2473. References 479 Drew, T., Võ, M.L.-H., & Wolfe, J. M. (2013). The invisible gorilla strikes again: Sustained inattentional blindness in expert observers. Psychological Science, 24, 1848–1853. Duchaine, B. C., & Nakayama, K. (2006). Developmental prosopagnosia: A window to content-specific face processing. Current Opinion in Neurobiology, 16, 166–173. Duffy, V. B., Lucchina, L. A., & Bartoshuk, L. M. (2004). Genetic variation in taste: Potential biomarker for cardiovascular disease risk? In J. Prescott & B. J. Tepper (Eds.), Genetic variation in taste sensitivity: Measurement, significance, and implications (pp. 195–228). New York, NY: Marcel Dekker. Ealy, G. T. (1994). Of ear trumpets and a resonance plate: Early hearing aids and Beethoven’s hearing perception. 19th-Century Music, 17, 262–273. Eggenberger, E. R. (2014). Supranuclear eye movement abnormalities. Continuum: Lifelong Learning in Neurology, 20, 981–992. Egly, R., Driver, J., & Rafal, R. D. (1994). Shifting visual attention between objects and locations: Evidence from normal and parietal lesion subjects. Journal of Experimental Psychology: General, 123, 161–177. Eimas, P. D., & Corbit, J. D. (1973). Selective adaptation of linguistic feature detectors. Cognitive Psychology, 4, 99–109. Engen, T. (1972). The effect of expectation on judgments of odor. Acta Psychologica, 36, 450–458. Enns, J. T., & Di Lollo, V. (1997). Object substitution: A new form of masking in unattended visual locations. Psychological Science, 8(2), 135–139. doi:10.1111/j.1467-9280.1997.tb00696.x Epstein, R. (2005). The cortical basis of visual scene processing. Visual Cognition, 12, 954–978. Epstein, R., & Kanwisher, N. (1998). A cortical representation of the local visual environment. Nature, 392, 598–601. Fain, G. L. (2003). Sensory transduction. Sunderland, MA: Sinauer. Farina, F. R., Mitchell, K. J., & Roche, R. A. P. (2017). Synaesthesia lost and found: Two cases of person- and music-colour synaesthesia. European Journal of Neuroscience, 45, 472–477. Fechner, G. T. (1966). Elements of psychophysics (H. E. Adler, Trans., D. H. Howes & E. G. Boring, Eds.). New York, NY: Holt, Rinehart. (Original work published 1860) Fedderson, W. E., Sandel, T. T., Teas, D. C., & Jeffress, L. A. (1957). Localization of high frequency tones. Journal of the Acoustical Society of America, 28, 988–991. Feynman, R. P. (1985). QED: The strange theory of light and matter. Princeton, NJ: Princeton University Press. Feynman, R. P. (1995). Six easy pieces: Essentials of physics explained by its most brilliant teacher. Reading, MA: Perseus. Fields, R. D. (2007). The shark’s electric sense. Scientific American, 297, 74–81. Findling, D. (2017). CNBC took BMW’s new self-driving car out for a spin. Here’s what it was like. CNBC News. Retrieved from http://www.cnbc.com/2017/01/13/cnbc- took-bmws-new-self-driving-car-out-for-a-spin-heres- what-it-was-like.html Firestone, C., & Scholl, B. J. (2016). Cognition does not affect perception: Evaluating the evidence for “top- down” effects. Behavioral and Brain Sciences, 39, e229. doi:10.1017/S0140525X15000965 Fisher, M. F. K. (1943). The gastronomical me. New York, NY: Duell, Sloan & Pearce. Foell, J., Bekrater-Bodmann, R., Diers, M., & Flor, H. (2014). Mirror therapy for phantom limb pain: Brain changes and the role of body representation. European Journal of Pain, 18(5), 729–739. Foldvari, M., & Chen, D.W. (2016). The intricacies of neurotrophic factor therapy for retinal ganglion cell rescue in glaucoma: A case for gene therapy. Neural Regeneration Research, 11, 875–877. Foster, D. H. (2011). Color constancy. Vision Research, 51, 674–700. Fowler, C. A., & Dekle, D. J. (1991). Listening with eye and hand: Cross-modal contributions to speech perception. Journal of Experimental Psychology: Human Perception and Performance, 17, 816–828. Fowler, C. A., & Thompson, J. M. (2010). Listeners’ perception of “compensatory shortening.” Attention, Perception, & Psychophysics, 72, 481–491. Frank, G. K. W., Oberndorfer, T. A., Simmons, A. N., Paulus, M. P., Fudge, J. L., Yang, T. T., & Kaye, W. H. (2008). Sucrose activates human taste pathways differently from artificial sweetener. NeuroImage, 39(4), 1559–1569. Fraser, A., & Wilcox, K. J. (1979). Perception of illusory movement. Nature, 281, 565–566. Frasnelli, J., Fark, T., Lehmann, J., Gerber, J., & Hummel, T. (2013). Brain structure is changed in congenital anosmia. Neuroimage, 83, 1074–1080. Frazier, L. D. (2018). Health psychology. New York, NY: Worth. Freeman, S., Yu, R., Egorova, N., Chen, X., Kirsch, I., Claggett, B., Kaptchuk, T. J., . . . Kong, J. (2015). Distinct neural representations of placebo and nocebo effects. Neuroimage, 112, 197–207. Froehlich, J. C. (1997). Opioid peptides. Alcohol Research and Health, 21(2), 132. Fromkin, V., Krashen, S., Curtiss, S., Rigler, D., & Rigler, M. (1974). The development of language in Genie: A case of language acquisition beyond the “critical period.” Brain and Language, 1(1), 81–107. Furl, N., Garrido, L., Dolan, R. J., Driver, J., & Duchaine, B. (2011). Fusiform gyrus face selectivity relates to Sensation and Perception480 individual differences in facial recognition ability. Journal of Cognitive Neuroscience, 23, 1723–1740. Galanter, E. (1962). Direct measurement of utility and subjective probability. American Journal of Psychology, 75, 208–220. Galantucci, B., Fowler, C. A., & Turvey, M. T. (2006). The motor theory of speech perception reviewed. Psychonomic Bulletin & Review, 13, 361–377. Galletti, C., & Fattori, P. (2003). Neuronal mechanisms for detection of motion in the field of view. Neuropsychologia, 41, 1717–1727. Gallivan, J. P., McLean, D. A., Flanagan, J. R., & Culham, J. C. (2013). Where one hand meets the other: Limb- specific and goal-dependent movement plans decoded from preparatory signals in single human parieto-frontal brain areas. Journal of Neuroscience, 33, 1991–2008. Gamito, P., Oliveira, J., Rosa, P., Morais, D., Duarte, N., Oliveira, S., & Saraiva, T. (2010). PTSD elderly war veterans: A clinical controlled pilot study. Cyberpsychology, Behavior, and Social Networking, 13(1), 43–48. Garneau, N. L., Nuessle, T. M., Sloan, M. M., Santorico, S. A., Coughlin, B. C., & Hayes J. E. (2014, May 27). Crowdsourcing taste research: Genetic and phenotypic predictors of bitter taste perception as a model. Frontiers in Integrative Neuroscience. Retrieved from http://journal.frontiersin.org/Journal/10.3389/ fnint.2014.00033/abstract Gatzia, D. E., & Schwartz, B. L. (2016). I think it’s going to hurt: The effects of cognition on pain perception. Presented at the meeting of the Southern Society of Philosophy and Psychology, March 12, 2016, Louisville, KY. Gelbard-Sagiv, H., Mukamel, R., Harel, M., Malach, R., & Fried, I. (2008). Internally generated reactivation of single neurons in human hippocampus during free recall. Science, 322, 96–101. Gelfand, S. A. (2009). Essentials of audiology (3rd ed.). New York, NY: Thieme Medical. Gerstmann, J. (2001). Pure tactile agnosia. Cognitive Neuropsychology, 18, 267–274. Gibson, E. J. (2001). Perceiving the affordances: A portrait of two psychologists. Hillsdale, NJ: Erlbaum. Gibson, J. J. (1950). The perception of the visual world. Boston, MA: Houghton Mifflin. Gibson, J. J. (1979). The ecological approach to visual perception. Hillsdale, NJ: Erlbaum. Gilbert, S. F. (2000). Developmental biology (6th ed.). Sunderland, MA: Sinauer Associates. Gilchrist, A., Kossyfidis, C., Bonato, F., Agostini, T., Cataliotti, J., Li, X., . . . Economou, E. (1999). An anchoring theory of lightness perception. Psychological Review, 106, 795–834. Godwin, H. J., Houpt, J. W., Walenchok, S. C., Hout, M. C., & Goldinger, S. D. (2015). Faster than the speed of rejection: Object identification processes during visual search for multiple targets. Journal of Experimental Psychology: Human Perception and Performance, 41, 1007–1020. Godwin, H. J., Menneer, T., Cave, K. R., Helman, S., Way, R. L., & Donnelly, N. (2010). The impact of relative prevalence on dual-target search for threat items from airport X-ray screening. Acta Psychologica, 134, 79–84. Goldberg, S., & Trattler, W. (2015). Ophthalmology made ridiculously simple. Miami, FL: Medmaster. Goldwater, S., Jurafsky, D., & Manning, C. D. (2010). Which words are hard to recognize? Prosodic, lexical, and disfluency factors that increase speech recognition error rates. Speech Communication, 52(3), 181–200. Goodale, M. A., Milner, A. D., Jakobsen, L. S., & Carey, D. P. (1991). A neurological dissociation between perceiving objects and grasping them. Nature, 349, 154–156. Graham, C. H., & Hsia, Y. (1958). Color defect and color theory. Science, 127, 675–682. Grant, P., Spencer, L., Arnoldussen, A., Hogle, R., Nau, A., Szlyk, J., . . . Seiple W. (2016). The functional performance of the brainport v100 device in persons who are profoundly blind. Journal of Visual Impairment & Blindness (Online), 110(2), 77. Granville, W. C. (1990). Colors do look different after a lens implant! Color Research & Application, 15(1), 59–62. Graven, S. N., & Browne, J. V. (2008). Auditory development in the fetus and infant. Newborn and Infant Nursing Review, 8, 187–193. Green, D. M., & Swets, J. (1966). Signal detection theory and psychophysics. New York, NY: Wiley. Green, E., & Murphy, C. (2012). Altered processing of sweet taste in the brain of diet soda drinkers. Physiology & Behavior, 107(4), 560–567. Greenwood, V. (2012, July–August). The humans with super human vision. Discover. Retrieved from http:// discovermagazine.com/2012/jul-aug/06-humans-with- super-human-vision Gregory, R. L. (1966). Eye and brain. New York, NY: McGraw-Hill. Grill-Spector, K., Knouf, N., & Kanwisher, N. (2004). The fusiform face area subserves face perception, not generic within-category identification. Nature Neuroscience, 7, 555–562. Grimault, N., Bacon, S. P., & Micheyl, C. (2002). Auditory stream segregation on the basis of amplitude-modulation rate. Journal of the Acoustical Society of America, 113, 1340–1348. Grossman, E. D., Battelli, L., & Pascual-Leone, A. (2005). Repetitive TMS over STS disrupts perception of biological motion. Vision Research, 45, 2847–2853. References 481 Grüsser, O.-J. (1993). The discovery of the psychophysical power law by Tobias Mayer in 1754 and the psychophysical hyperbolic law by Ewald Hering in 1874. Behavioral and Brain Sciences, 16, 142–144. Guyton, A. C. (1991). Textbook of medical physiology (8th ed.). Philadelphia, PA: Saunders. Habibi, A., Cahn, R., Damasio, A., & Damasio, H. (2016). Neural correlates of accelerated auditory processing in children engaged in music training. Developmental Cognitive Neuroscience, 21, 1–14. Habibi, A., Wirantana, V., & Starr, A. (2013). Cortical activity during perception of musical pitch: Comparing musicians and nonmusicians. Music Perception, 30, 463–479. Hahnel, U. J. J., & Hecht, H. (2012). The impact of rear-view mirror distance and curvature on judgements relevant to road safety. Ergonomics, 55, 23–26. Hansen, J. H., & Arslan, L. M. (1995, May). Foreign accent classification using source generator based prosodic features. In IEEE International Conference on Acoustics, Speech, and Signal Processing, 1995 (Vol. 1, pp. 836–839). Hansen, T., Olkkonen, M., Walter, S., & Gegenfurtner, K. R. (2006). Memory modulates color appearance. Nature Neuroscience, 9(11), 1367–1368. Harley, T. A. (2014). The psychology of language: From data to theory (4th ed.). New York, NY: Psychology Press. Hartline, H. K., & Ratliff, F. (1958). Spatial summation of inhibitory influences in the eye of Limulus and the mutual interaction of receptor units. Journal of General Physiology, 41, 1049–1066. Hartong, D. T., Berson, E. L., & Dryja, T. P. (2006). Retinitis pigmentosa. Lancet, 368, 1795–1809. Håstad, O., Victorsson, J., & Ödeen, A. (2005). Differences in color vision make passerines less conspicuous in the eyes of their predators. PNAS, 102(18), 6391–6394. Hastorf, A. H., & Cantril, H. (1954). They saw a game: A case study. Journal of Abnormal and Social Psychology, 49, 129–134. He, J. (2003). Corticofugal modulation of the auditory thalamus. Experimental Brain Research, 153, 579–590. Hecht, H., & Savelsbergh, G. J. P. (2004). Time-to-contact. Amsterdam, The Netherlands: Elsevier. Hecht, S., Schlaer, S., & Pirenne, M. H. (1942). Energy, quanta, and vision. Journal of General Physiology, 39, 651–673. Hedges, J. H., Gartshteyn, Y., Kohn, A., Rust, N. C., Shadlen, M. N., Newsome, W. T., & Movshon, J. A. (2011). Dissociation of neuronal and psychophysical responses to local and global motion. Current Biology, 21, 2023–2028. Held, R. (1965). Plasticity in sensory-motor systems. Scientific American, 213, 84–94. Held, R. (1980). The rediscovery of adaptability in the visual system: Effects of extrinsic and intrinsic chromatic dispersion. In C. S. Harris (Ed.), Visual coding and adaptability (pp. 69–94). Hillsdale, NJ: Erlbaum. Held, R., & Hein, A. (1963). Movement-produced stimulation in the development of visually guided behavior. Journal of Comparative and Physiological Psychology, 56, 872–876. Heller, M. A., Brackett, D. D., Wilson, K., Yoneyama, K., Boyer, A., & Steffen, H. (2002). The haptic Müller-Lyer illusion in sighted and blind people. Perception, 31(10), 1263–1274. Hendry, S. H. C., & Reid, R. C. (2000). The koniocellular pathway in primate vision. Annual Review of Neuroscience, 23, 127–153. Henry, B. A., Turner, C. W., & Behrens, A. (2005). Spectral peak resolution and speech recognition in quiet: Normal hearing, hearing impaired, and cochlear implant listeners. Journal of the Acoustical Society of America, 118, 1111–1121. Hershenson, M., & Samuels, S. M. (1999). An airplane illusion: Apparent velocity determined by apparent distance. Perception, 28, 433–436. Herz, R. (2007). The scent of desire. New York, NY: HarperCollins. Herz, R. S., & von Clef, J. (2001). The influence of verbal labelling on the perception of odors: Evidence for olfactory illusions? Perception, 30, 381–391. Herz, R. S., Eliassen, J. C., Beland, S. L., & Souza, T. (2004). Neuroimaging evidence for the emotional potency of odor-evoked memory. Neuropsychologia, 42, 371–378. Hickok, G., & Poeppel, D. (2007). The cortical organization of speech processing. Nature Reviews Neuroscience, 8, 393–401. Hof, P. R., & van der Gucht, E. (2007). Structure of the cerebral cortex of the humpback whale, Megaptera novaeangliae (Cetacea, Mysticeti, Balaenopteridae). Anatomical Record, 290, 1–31. Hofeldt, A. J., & Hoefle, F. B. (1993). Stereophotometric testing for Pulfrich’s phenomenon in professional baseball players. Perceptual and Motor Skills, 77, 407–416. Hofeldt, A. J., Hoefle, F. B., & Bonafede, B. (1996). Baseball hitting, binocular vision, and the Pulfrich phenomenon. Archives of Ophthalmology, 114, 1490–1494. Hoffman, P. M., Van Riswick, J. G., & Van Opstal, A. J. (1998). Relearning sound localization with new ears. Nature Neuroscience, 1, 417–421. Hollins, M., Bensmaia, S., & Washburn, S. (2001). Vibrotactile adaptation impairs discrimination of fine, but not coarse, textures. Somatosensory and Motor Research, 18, 253–262. Sensation and Perception482 Holt, L. L. (2005). Temporally non-adjacent non-linguistic sounds affect speech categorization. Psychological Science, 16, 305–312. Holt, L. L., & Lotto, A. J. (2010). Speech perception as categorization. Attention, Perception, & Psychophysics, 72, 1218–1227. Holway, A. H., & Boring, E. G. (1941). Determinants of apparent visual size with distance variant. American Journal of Psychology, 54, 21–37. Horridge, A. (2012). Visual discrimination by the honeybee (Apis mellijera). In O. F. Lazareva, T. Shimizu, & E. A. Wasserman (Eds.), How animals see the world: Comparative behavior, biology, and evolution of vision (pp. 165–190). New York, NY: Oxford University Press. Horswill, M., Helman, S., Ardiles, P., & Wann, J. P. (2005). Motorcycle accident risk could be inflated by a time to arrival illusion. Optometry and Vision Science, 82, 740–746. Howard, I. P., & Rogers, B. J. (2002). Seeing in depth: Vol. 2. Depth perception. Toronto, Canada: Porteus. Howard, J. D., Plailly, J., Grueschow, M., Haynes, J. D., & Gottfried, J. A. (2009). Odor quality coding and categorization in human posterior piriform cortex. Nature Neuroscience, 12, 932–938. Hsiao, S. (2008). Central mechanisms of tactile shape perception. Current Opinion in Neurobiology, 18, 418–424. Hua, L. H., Strigo, I. A., Baxter, L. C., Johnson, S. C., & Craig, A. D. (2005). Anteroposterior somatotopy of innocuous cooling activation focus in human dorsal posterior insular cortex. American Journal of Physiology: Regulatory, Integrative, and Comparative Physiology, 289, R319–R325. Huang, Q., & Tang, J. (2010). Age-related hearing loss or presbycusis. European Archives of Oto-Rhino- Laryngology, 267(8), 1179–1191. Hubbard, E. M., Brang, D., & Ramachandran, V. S. (2011). The cross-activation theory at 10. Journal of Neuropsychology, 5, 155–177. Hubel, D. H., & Wiesel, T. N. (1959). Receptive fields of single neurons in the cat’s striate cortex. Journal of Physiology, 148, 574–591. Hubel, D. H., & Wiesel, T. N. (1962). Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. Journal of Physiology, 160, 106–154. Hubel, D. H., & Wiesel, T. N. (1965). Receptive fields and functional architecture in two nonstriate visual areas (18 and 19) of the cat. Journal of Neurophysiology, 28, 229–289. Hubel, D. H., & Wiesel, T. N. (1979). Brain mechanisms and vision. Scientific American, 241, 150–162. Hubel, D. H., & Wiesel, T. N. (2005). Brain and visual perception: The story of a 25-year collaboration. New York, NY: Oxford University Press. Hughes, H. C. (1999). Sensory exotica: A world beyond experience. Cambridge, MA: MIT Press. Hughes, H. C., & Zimba, L. D. (1985). Spatial maps of directed visual attention. Journal of Experimental Psychology: Human Perception and Performance, 11, 409–430. Humayun, M. S., Dorn, J. D., da Cruz, L., Dagnelie, G., Sahel, J. A., Stanga, P. E., . . . Greenberg, R. J. Argus II Study Group. (2012). Interim results from the international trial of Second Sight’s visual prosthesis. Ophthalmology, 119, 779–788. Hurvich, L. M., & Jameson, D. (1957). An opponent-process theory of color vision. Psychological Review, 64, 384–404. Hutchinson, W. D., Davis, K. D., Lozano, A. M., Tasker, R. R., & Dostrovsky, J. O. (1999). Pain-related neurons in the human cingulate cortex. Nature Neuroscience, 2, 403–405. Huttenlocher, P. R., & de Courten, C. (1987). The development of synapses in striate cortex of man. Human Neurobiology, 6, 1–9. Hyde, K., Peretz, I., & Zatorre, R. (2008). Evidence for the role of the right auditory cortex in fine pitch resolution. Neuropsychologia, 46, 632–639. Irons, J., & Remington, R. (2013). Can attentional control settings be maintained for two color-location conjunctions? Evidence from an RSVP task. Attention, Perception, & Psychophysics, 75, 862–875. Isaacs, L. D. (1981). Relationship between depth perception and basketball-shooting performance over a competitive season. Perceptual and Motor Skills, 53, 554. Jacquin-Cortois, S., O’Shea, J., Luaté, J., Pisella, L., Revol, P., . . . Rossetti, Y. (2013). Rehabilitation of spatial neglect by prism adaptation: A peculiar expansion of sensorimotor after-effects to spatial cognition. Neuroscience and Biobehavioral Reviews, 37, 594–609. James, W. (1890). The principles of psychology. New York, NY: Holt. Jameson, K. A., Highnote, S., & Wasserman, L. (2001). Richer color experience in observers with multiple photopigment opsin genes. Psychonomic Bulletin & Review, 8, 244–261. Janata, P., Birk, J. L., Van Horn, J. D., Leman, M., Tillmann, B., & Bharucha, J. J. (2002). The cortical topography of tonal structures underlying Western music. Science, 298, 2167–2170. Johansson, G. (1973). Visual perception of biological motion and a model for its analysis. Perception and Psychophysics, 14, 195–204. Johnson, E. N., Hawken, M. J., & Shapley, R. (2001). The spatial transformation of color in the primary visual cortex of the macaque monkey. Nature Neuroscience, 4, 409–413. Johnsrude, I. S., Mackey, A., Hakyemez, H., Alexander, E., Trang, H. P., & Carylon, R. P. (2013). Swinging at a References 483 cocktail party: Voice familiarity aids speech perception in the presence of a competing voice. Psychological Science, 24, 1995–2004. Jones, F. N. (1956). A forced-choice method of limits. American Journal of Psychology, 69, 672–673. Jönsson, F. U., & Olsson, M. J. (2012). Knowing what we smell. In G. M. Zucco, R. S. Herz, & B. Schaal (Eds.), Olfactory cognition: From perception and memory to environmental odours and neuroscience (pp. 115–136). Amsterdam, The Netherlands: John Benjamins. Jönsson, F. U., & Stevenson, R. J. (2014). Odor knowledge, odor naming and the “tip of the nose” experience. In B. L. Schwartz & A. S. Brown (Eds.), Tip-of-the- tongue states and related phenomena (pp. 305–326). Cambridge, UK: Cambridge University Press. Jordan, G., Deeb, S. S., Bosten, J. M., & Mollon, J. D. (2010). The dimensionality of color vision in carriers of anomalous trichromacy. Journal of Vision, 10, 1–19. Juang, B. H., & Rabiner, L. R. (2005). Automatic speech recognition: A brief history of the technology development. Georgia Institute of Technology. Atlanta Rutgers University and the University of California. Santa Barbara, 1, 67. Julesz, B. (1971). Foundations of cyclopean perception. Chicago, IL: University of Chicago Press. Kaas, A., Weigelt, S., Roebroeck, A., Kohler, A., & Muckli, L. (2010). Imagery of a moving object: The role of occipital cortex and human MT/V5+. Neuroimage, 49, 794–804. Kadohisa, M., & Wilson, D. A. (2006). Separate encoding of identity and similarity of complex familiar odors in piriform cortex. Proceedings of the National Academy of Sciences of the United States of America, 103, 15206–15211. Kallus, K. W., Tropper, K., & Boucsein, W. (2011). The importance of motion cues in spatial disorientation training for VFR-pilots. International Journal of Aviation Psychology, 21, 135–152. Kandil, F. I., Rotter, A., & Lappe, M. (2010). Car drivers attend to different gaze targets when negotiating closed vs. open bends. Journal of Vision, 10, 1–11. Kanizsa, G. (1979). Organization in vision: Essays on gestalt perception. New York, NY: Praeger. Kanwisher, N., & Dilks, D. (2013). The functional organization of the ventral visual pathway in humans. In L. Chalupa & J. Werner (Eds.), The new visual neurosciences (pp. 733–748). Cambridge, MA: MIT Press. Kanwisher, N., McDermott, J., & Chun, M. M. (1997). The fusiform face area: A module in human extrastriate cortex specialized for face perception. Journal of Neuroscience, 17, 4302–4311. Kanwisher, N., & Yovel, G. (2006). The fusiform face area: A cortical region specialized for the perception of faces. Philosophical Transactions of the Royal Society of London, Series B: Biological Sciences, 361, 2109–2128. Kapoula, Z., Bernotas, M., & Haslwanter, T. (1999). Listing’s plane rotation with convergence: Role of disparity, accommodation, and depth perception. Experimental Brain Research, 126(2), 175–186. Kau, S., Strumpf, H., Merkel, C., Stoppel, C. M., Heinze, H. J., Hopf, J. M., & Schoenfeld, M. A. (2013). Distinct neural correlates of attending speed vs. coherence of motion. Neuroimage, 64, 299–307. Kawahara, J. I., & Sato, H. (2013). The effect of fatigue on the attentional blink. Attention, Perception, & Psychophysics, 75, 1096–1102. Keane, A. G. (2011, December 8). U.S. highway deaths decline for a fifth year, longest streak since 1899. Bloomberg News. Retrieved from https://www .bloomberg.com/news/articles/2011-12-08/u-s-highway- deaths-decline-2-9-falling-for-fifth-year-1- Kirchhof, H. (1950). A method for the objective measurement of accommodation speed of the human eye. American Journal of Optometry, 27, 163–178. Kitaoka, A. (2005). Trick eyes graphics. Tokyo, Japan: Kanzen. Klatzky, R. L., Loomis, J. M., Lederman, S. J., Wake, H., & Fujita, N. (1993). Haptic identification of objects and their depictions. Perception and Psychophysics, 54, 170–178. Klüver, H., & Bucy, P. C. (1939). Preliminary analysis of function of the temporal lobes in monkeys. Archives of Neurology and Psychiatry, 42, 979–1000. Kochkin, S. (2005). MarkeTrak VII: Customer satisfaction with hearing instruments in the digital age. Hearing Journal, 58, 30–44. Kooijman, C. M., Dijkstra, P. U., Geertzen, J. H. B., Elzinga, A., & van der Schans, C. P. (2000). Phantom pain and phantom sensations in upper limb amputees: An epidemiological study. Pain, 87, 33–41. Krantz, J. H., Silverstein, L. D., & Yeh, Y.-Y. (1992). Visibility of transmissive liquid crystal displays under dynamic lighting conditions. Human Factors, 34, 615–632. Krings, T., Töpper, R., Foltys, H., Erberich, S., Sparing, R., Willmes, K., & Thron, A. (2000). Cortical activation patterns during complex motor tasks in piano players and control subjects: A functional magnetic resonance imaging study. Neuroscience Letters, 278, 189–193. Kristjánsson, A. (2013). The case for causal influences of action videogame play upon vision and attention. Attention, Perception, & Psychophysics, 75, 667–672. Kuffler, S. W. (1953). Discharge patterns and functional organization of mammalian retina. Journal of Neurophysiology, 16, 37–68. LaMotte, R. H., Shimada, S. G., Green, B. G., & Zelterman, D. (2009). Pruritic and nociceptive sensations and Sensation and Perception484 dysesthesias from a spicule of cowhage. Journal of Neurophysiology, 101, 1430–1443. Land, E. H. (1977). The retinex theory of color vision. Scientific American, 237, 108–128. Landis, B. N., Croy, I., & Haehner, A. (2012). Long lasting phantosmia treated with venlafaxine. Neurocase, 18, 112–114. Landry, S. P., Shiller, D. M., & Champoux, F. (2013). Short- term visual deprivation improves the perception of harmonicity. Journal of Experimental Psychology: Human Perception and Performance, 39, 1503–1507. Langenheim, J. (2010, September 17). The last of the sea nomads. The Guardian. Retrieved from http://www .theguardian.com/environment/2010/sep/18/last-sea- nomads Larsen, A., Madsen, K. H., Lund, T. E., & Bundesen, C. (2006). Images of illusory motion in primary visual cortex. Journal of Cognitive Neuroscience, 18, 1174–1180. Lawless, H. T. (1991). A sequential contrast effect in odor perception. Bulletin of the Psychonomic Society, 29, 317–319. Lederman, S. J., & Klatzky, R. L. (2009). Haptic perception: A tutorial. Attention, Perception, & Psychophysics, 71, 1439–1459. Leek, M. R., & Molis, M. R. (2009). Beyond audibility: Hearing loss and the perception of speech. ASHA Leader, 14, 14–17. Lettvin, J. Y., Maturana, H. R., McCulloch, W. S., & Pitts, W. H. (1968). What the frog’s eye tells the frog’s brain. In W. C. Corning & M. Baladan (Eds.), The mind: Biological approaches to its functions (pp. 233–258). New York, NY: Interscience. Li, W., Lopez, L., Osher, J., & Howard, J. D. (2010). Right orbitofrontal cortex mediates conscious olfactory perception. Psychological Science, 21, 1454–1463. Liang, M., Mouraux, A., Hu, L., & Iannetti, G. D. (2013). Primary sensory cortices contain distinguishable spatial patterns of activity for each sense. Nature Communications, 4, 1979. Liberman, A. M., & Mattingly, I. G. (1985). The motor theory of speech perception. Cognition, 21, 1–36. Limb, C. J., & Roy, A. T. (2014). Technological, biological, and acoustical constraints to music perception in cochlear implant users. Hearing Research, 308, 13–26. Liu, J., Harris, A., & Kanwisher, N. (2010). Perception of face parts and face configurations: An fMRI study. Journal of Cognitive Neuroscience, 22, 203–211. Livadas, C., Lygeros, J., & Lynch, N. A. (2000). High- level modeling and analysis of the traffic alert and collision avoidance system (TCAS). Proceedings of the IEEE, 88(7), 926–948. Livingstone, M. (2002). Vision and art: The biology of seeing. New York, NY: Abrams. Livingstone, M. S., & Hubel, D. H. (1984). Anatomy and physiology of a color system in the primate visual cortex. Journal of Neuroscience, 4, 309–356. Lokki, T., & Patynen, J. (2015). The acoustics of a concert hall as a linear problem. Europhysics News, 46, 15–18. Looser, C. E., & Wheatley, T. (2010). The tipping point of animacy: How, when, and where we perceive life in a face. Psychological Science, 21, 1854–1862. Loui, P., Zamm, A., & Schlaug, G. (2012). Absolute pitch and synesthesia: Two sides of the same coin? Shared and distinct neural substrates of music listening. Paper presented at the 12th International Conference for Music Perception and Cognition, Thessaloniki, Greece. Lowenfield, B., & Abel, G. L. (1977). Methods of teaching Braille reading efficiency of children in lower senior classes. Birmingham, UK: Research Centre for the Education of the Visually Handicapped. Mack, A., & Rock, I. (1998). Inattentional blindness. Cambridge, MA: MIT Press. MacMahon, M. K. C. (1996). Phonetic notation. In P. T. Daniels & W. Bright (Eds.), The world’s writing systems (pp. 821–846). New York, NY: Oxford University Press. Maertens, M., & Pollmann, S. (2005). fMRI reveals a common neural substrate of illusory and real contours in V1 after perceptual learning. Journal of Cognitive Neuroscience, 17, 1553–1564. Malnic, B., Godfrey, P. A., & Buck, L. B. (2004). The human olfactory receptor gene family. Proceedings of the National Academy of Sciences of the United States of America, 101, 2584–2589. Marchi, F., & Newen, A. (2015). Cognitive penetrability and emotion recognition in human facial expressions. Frontiers in Psychology, 6, 828. doi:10.3389/ fpsyg.2015.00828 Marr, D. (1982). Vision. San Francisco, CA: W. H. Freeman. Masland, R. H. (2012). The tasks of amacrine cells. Visual Neuroscience, 29, 3–9. Mason, C., & Kandel, E. R. (1991). Central visual pathways. In E. R. Kandel, J. H. Schwartz, & T. M. Jessel (Eds.), Principles of neural science (3rd ed., pp. 420–439). Norwalk, CT: Appleton & Lange. Mattys, S. L., Barden, K., & Samuel, A. G. (2014). Extrinsic cognitive load impairs low-level speech perception. Psychonomic Bulletin & Review, 21, 748–754. Maurer, D., & Werker, J. F. (2014). Perceptual narrowing during infancy: A comparison of language and faces. Developmental Psychobiology, 56, 154–178. May, P. J. (2006). The mammalian superior colliculus: Laminar structure and connections. Progress in Brain Research, 151, 321–378. Mayer, D. L., & Dobson, V. (1982). Visual acuity development in infants and young children, as assessed References 485 by operant preferential looking. Vision Research, 22(9), 1141–1151. Mazyn, L. I. N., Lenoir, M., Montagne, G., & Savelsbergh, G. J. (2004). The contribution of stereo vision to one- handed catching. Experimental Brain Research, 157, 383–390. McCarley, J. S., Kramer, A. F., Wickens, C. D., Vidoni, E. D., & Boot, W. R. (2004). Visual skills in airport- security screening. Psychological Science, 15, 302–306. McGurk, H., & MacDonald, T. (1976). Hearing lips and seeing voices. Nature, 264, 746–748. McKeegan, N. (2013, February 18). Argus II becomes first “bionic eye” to gain approval for sale in the U.S. New Atlas. Retrieved from http://newatlas.com/argus-ii- becomes-first-bionic-eye-to-gain-approval-for-sale-in- us/26295 Melzack, R., & Wall, P. D. (1988). The challenge of pain (2nd ed.). New York, NY: Penguin. Menashe, I., Man, O., Lancet, D., & Gilad, Y. (2003). Different noses for different people. Nature Genetics, 34, 143–144. Mendez, M. F., & Cherrier, M. M. (2003). Agnosia for scenes in topographagnosia. Neuropsychologia, 41, 1387–1395. Meng, M., Remus, D. R., & Tong, F. (2005). Filling-in of visual phantoms in the human brain. Nature Neuroscience, 8, 1248–1254. Menneer, T., Stroud, M., Cave, K. R., Donnelly, N., & Rayner, K. (2008). Eye movements in search for multiple targets. In K. Rayner, D. Shen, X. Bai, & G. Yan (Eds.), Cognitive and cultural influences on eye movements (pp. 197– 208). London, UK: Psychology Press. Meyer, R. (2009). Acoustics and the performance of music. New York, NY: Springer. Meyer, R., Palmer, C., & Mazo, M. (1998). Affective and coherence responses to Russian laments. Music Perception, 16, 135–150. Mignault-Goulet, G., Moreau, P., Robitaille, N., & Peretz, I. (2012). Congenital amusia persists in the developing brain after daily music listening. PLoS ONE, 7, e36860. Miller, L. R., & Miller, F. G. (2015). Understanding placebo effects: Implications for nursing practice. Nursing Outlook, 63(5), 601–606. doi:10.1016/j .outlook.2015.03.002 Misaka, T. (2013). Molecular mechanisms of the action of miraculin, a taste-modifying protein. Seminars in Cell & Developmental Biology, 24, 222–225. Mishkin, M., Ungerleider, L. G., & Macko, K. A. (1983). Object vision and spatial vision: Two cortical pathways. Trends in Neurosciences, 6, 414–417. Møller, J. (2006). Hearing: Anatomy, physiology, and disorders of the auditory system (2nd ed.). Amsterdam, The Netherlands: Elsevier. Mollon, J. (1992). Worlds of difference. Nature, 356, 378–379. Monnier, P. (2008). Standard definitions of chromatic induction fail to describe induction with S-cone patterned backgrounds. Vision Research, 48, 2708–2714. Moore, B. C. J. (2012). Effects of bandwidth, compression speed, and gain at high frequencies on preferences for amplified music. Trends in Amplification, 16, 159–172. Moran, J., & Desimone, R. (1985). Selective attention gates visual processing in extrastriate cortex. Science, 229, 782–784. Moreau, P., Jolicoeur, P., & Peretz, I. (2013). Pitch discrimination without awareness in congenital amusia: Evidence from event-related potentials. Brain and Cognition, 81, 337–344. Morgan, M. J., Adam, A., & Mollon, J. D. (1992). Dichromats detect colour-camouflaged objects that are not detected by trichromats. Proceedings of the Royal Society of London, Series B: Biological Sciences, 248(1323), 291–295. Moscovitch, M., & Moscovitch, D. (2000). Super face- inversion effects for isolated internal or external features, and for fractured faces. Cognitive Neuropsychology, 17, 201–219. Mountcastle, V. B. (2005). The sensory hand: Neural mechanisms of somatic sensation. Cambridge, MA: Harvard University Press. Murphy, G., & Kovach, J. K. (1972). Historical introduction to modern psychology (3rd ed.). New York, NY: Harcourt Brace. Nagata, T., Koyanagi, M., Tsukamoto, H., Saeki, S., Isono, K., Shichida, Y., . . . Terakita, A. (2012). Depth perception from image defocus in a jumping spider. Science, 335, 469–471. Nagayama, S., Takahashi, Y. K., Yoshihara, Y., & Mori, K. (2004). Mitral and tufted cells differ in the decoding manner of odor maps in the rat olfactory bulb. Journal of Neurophysiology, 91, 2532–2540. Nassi, J. J., & Callaway, E. M. (2009). Parallel processing strategies of the primate visual system. Nature Reviews Neuroscience, 10, 360–372. Nathans, J., Piantanida, T., Eddy, R., Shows, T., & Hogness, D. (1986). Molecular genetics of inherited variation in human color vision. Science, 232, 203–210. National Institute on Deafness and Other Communication Disorders. (2016). Cochlear implants (NIH Publication No. 11-4798). Bethesda, MD: Author. National Institutes of Health. (n.d.). Hearing loss: A common problem for older adults. Retrieved from http:// nihseniorhealth.gov/hearingloss/hearinglossdefined/01 .html Neisser, U. (1979). The control of information pickup in selective looking. In A. D. Pick (Ed.), Perception and its development: A tribute to Eleanor Gibson (pp. 201–219). Hillsdale, NJ: Erlbaum. Sensation and Perception486 Neitz, J., & Jacobs, G. (1986). Polymorphism of the long- wavelength cone in normal human colour vision. Nature, 323, 623–625. Neitz, J., & Jacobs, G. (1990). Polymorphism in normal human color vision and its mechanism. Vision Research, 30, 620–636. Neumeyer, C., & Arnold, K. (1989). Tetrachromatic color vision in the goldfish becomes trichromatic under white adaptation light of moderate intensity. Vision Research, 29(12), 1719–1727. Newsome, W. T., Britten, K. H., & Movshon, J. A. (1989). Neuronal correlates of a perceptual decision. Nature, 341, 52–54. Newsome, W. T., & Paré, E. B. (1988). A selective impairment of motion perception following lesions of the middle temporal visual area (MT). Journal of Neuroscience, 8, 2201–2211. Newsome, W. T., Shadlen, M. N., Zohary, E., Britten, K. H., & Movshon, J. A. (1995). Visual motion: Linking neuronal activity to psychophysical performance. In M. S. Gazzaniga (Ed.), The cognitive neurosciences (pp. 401–414). Cambridge, MA: MIT Press. Nilsson, D.-E. (1989). Vision, optics, and evolution. Bioscience, 39(5), 298–307. Nodine, C. F., Krupinski, E. A., & Kundel, H. L. (1993). Visual processing and decision making in search and recognition of targets embedded in pictorial scenes. In D. Brogan & A. G. Gale (Eds.), Visual search (pp. 239–249). Philadelphia, PA: Taylor & Francis. Nooij, S. A. E., & Groen, E. L. (2011). Rolling into spatial disorientation: Simulator demonstration of the post-roll (Gillingham) illusion. Aviation, Space, and Environmental Medicine, 82, 505–512. Olausson, H., Wessberg, J., Morrison, I., McGlone, F., & Valbo, A. (2010). The neurophysiology of unmyelinated tactile afferents. Neuroscience and Behavioral Reviews, 34, 185–191. Oltman, M. (2009). Migraines, perfume, and the Americans with Disabilities Act. Retrieved from http://www .healthcentral.com/migraine/c/11175/67623/disabilities-act Opriş, D., Pintea, S., García-Palacios, A., Botella, C., Szamosközi, Ş., & David, D. (2012). Virtual reality exposure therapy in anxiety disorders: A quantitative meta-analysis. Depression and Anxiety, 29(2), 85–93. Orban, G. A., Kennedy, H., & Butler, J. (1986). Velocity sensitivity and direction sensitivity of neurons in areas V1 and V2 of monkeys: Influence of eccentricity. Journal of Neurophysiology, 56, 462–480. Otero-Millan, J., Macknik, S. L., & Martinez-Conde, S. (2012). Microsaccades and blinks trigger illusory rotation in the “rotating snakes” illusion. Journal of Neuroscience, 32, 6043–6051. Overy, K., Peretz, I., Zatorre, R., Lopez, L., & Majno, M. (Eds.). (2012, April). The neurosciences and music IV: Learning and memory [Special issue]. Annals of the New York Academy of Sciences, 1252. Parker, A. J. (2007). Binocular depth perception and the cerebral cortex. Nature Reviews Neuroscience, 8, 379–391. Parkinson, C., Walker, T., Memmi, S., & Wheatley T. (2017). Emotions are understood from biological motion across remote cultures. Emotion, 16, 459–477. Partanen, E., Kujalaa, T., Näätänen, R., Litola, A., Sambeth, A., & Huotilainen, M. (2013). Learning-induced neural plasticity of speech processing before birth. Proceedings of the National Academy of Sciences of the United States of America, 110, 15145–15150. Pasley, B., David, S. V., Mesgarani, N., Flinker, A., Shamma, S. A., Crone, N. E., . . . Chang, E. F. (2012). Reconstructing speech from human auditory cortex. PLoS Biology, 10, 1–13. Pasternak, T., Bisley, J. W., & Calkins, D. (2003). Visual processing in the primate brain. In M. Gallagher & R. J. Nelson (Eds.), Handbook of psychology: Biological psychology (Vol. 3, pp. 139–185). Hoboken, NJ: Wiley & Sons. Pasupathy, A., & Connor, C. E. (2002). Population coding of shape in V4. Nature Neuroscience, 5, 1332–1338. Patel, A. D. (2008). Music, language, and the brain. New York, NY: Oxford University Press. Patel, A. D. (2013). Sharing and nonsharing of brain resources for language and music. In M. Arbib (Ed.), Language, music, and the brain (pp. 329–355). Cambridge, MA: MIT Press. Patynen, J., & Lokki, T. (2016). Perception of music dynamics in concert hall acoustics. Journal of the Acoustical Society of America, 140, 3787–3798. Pavani, F., Spence, C., & Driver, J. (2000). Visual capture of touch: Out-of-the-body experiences with rubber gloves. Psychological Science, 11(5), 353–359. Pegna, A. J., Khateb, A., Lazeyras, F., & Seghier, M. (2005). Discriminating emotional faces without the primary visual cortices involves the right amygdala. Nature Neuroscience, 8, 24–25. Pei, Y. C., Hsiao, S. S., Craig, J. C., & Bensmaia, S. J. (2011). Neural mechanisms of tactile motion integration in somatosensory cortex. Neuron, 69, 536–547. Penfield, W., & Rasmussen, T. (1950). The cerebral cortex of man: A clinical study of localization and function. New York, NY: Macmillan. Peretz, I. (2008). Musical disorders: From behavior to genes. Current Directions in Psychological Science, 17, 329–333. Peretz, I. (2013). The biological foundations of music: Insights from congenital amusia. In D. Deutsch (Ed.), References 487 The psychology of music (pp. 551–564). New York, NY: Elsevier. Peretz, I., & Hyde, K. L. (2003). What is specific to music processing? Insights from congenital amusia. Trends in Cognitive Sciences, 7, 362–367. Perlman, M., & Krumhansl, C. L. (1996). An experimental study of internal interval standards in Javanese and Western musicians. Music Perception, 14, 95–116. Perry, B. N., Mercier, C., Pettifer, S. R., Cole, J., & Tsao, J. W. (2014). Virtual reality therapies for phantom limb pain. European Journal of Pain, 18(7), 897–899. Peterson, M. A., & Salvagio, E. (2008). Inhibitory competition in figure-ground perception: Context and convexity. Journal of Vision, 8, 1–13. Petry, S., & Meyer, G. E. (1987). The perception of illusory contours. New York, NY: Springer. Pettigrew, J. D. (1999). Electroreception in monotremes. Journal of Experimental Biology, 202, 1447–1454. Pinckers, A. J. L. G. (1980). Color vision and age. Ophthalmologica, 181(1), 23–30. Pinker, S. (1994). The language instinct. New York, NY: William Morrow. Plantinga, J., & Trainor, L. J. (2005). Memory for melody: Infants use a relative pitch code. Cognition, 98, 1–11. Posner, M. I. (1980). Orienting of attention. Quarterly Journal of Experimental Psychology, 32, 3–25. Posner, M. I., & Rothbart, M. K. (2013). Development of attention networks. In B. R. Kar (Ed.), Cognition and brain development: Converging evidence from various methodologies (pp. 61–83). Washington, DC: American Psychological Association. Proctor, R. W., & Van Zandt, T. (2008). Human factors in simple and complex systems (2nd ed.). Boca Raton, FL: CRC Press. Ptito, A., & Leh, S. E. (2007). Neural substrates of blindsight after hemispherectomy. Neuroscientist, 13, 506–518. Quiroga, R. Q., Reddy, L., Kreiman, G., Koch, C., & Fried, I. (2005). Invariant visual representation by single neurons in the human brain. Nature, 435, 1102–1107. Quiroga, R. Q., Reddy, L., Kreiman, G., Koch, C., & Fried, I. (2008). Sparse but not “grandmother-cell” coding in the medial temporal lobe. Trends in Cognitive Sciences, 12, 87–91. Quiroga, R. Q., Snyder, L. H., Batista, A. P., Cui, H., & Andersen, R. A. (2006). Movement intention is better predicted than attention in the posterior parietal cortex. Journal of Neuroscience, 26, 3615–3620. Radtke, N. D., Aramant, R. B., Petry, H. M., Green, P. T., Pidwell, D. J., & Seiler, M. J. (2008). Vision improvement in retinal degeneration patients by implantation of retina together with retinal pigment epithelium. American Journal of Ophthalmology, 146, 177–182. Rainville, P., Duncan, G. H., Price, D. D., Carrier, B., & Bushnell, M. C. (1997). Pain affect encoded in human anterior cingulate, but not somatosensory cortex. Science, 277, 968–971. Ramachandran, V. S. (1992). Filling in the blind spot. Nature, 356, 115. Ramachandran, V. S., & Hirstein, W. (1998). The perception of phantom limbs: The D. O. Hebb lecture. Brain, 121, 1603–1630. Rauschecker, J. P., & Scott, S. K. (2009). Maps and streams in the auditory cortex: Nonhuman primates illuminate human speech processing. Nature Neurosciences, 12, 718–724. Raymond, J. E., Shapiro, K. L., & Arnell, K. M. (1992). Temporary suppression of visual processing in an RSVP task: An attentional blink? Journal of Experimental Psychology: Human Perception and Performance, 18(3), 849–860. Read, J. C. A. (2005). Early computational processing in binocular vision and depth perception. Progress in Biophysics and Molecular Biology, 87, 77–108. Reed, C. L., Caselli, R. J., & Farah, M. J. (1996). Tactile agnosia: Underlying impairment and implications for normal tactile object recognition. Brain, 119, 875–888. Reichardt, W. (1969). Movement perception in insects. In W. Reichardt (Ed.), Processing of optical data by organisms and by machines. New York, NY: Academic Press. Rensink, R. A., O’Regan, J. K., & Clark, J. (1997). To see or not to see: The need for attention to perceive changes in scenes. Psychological Science, 8, 368–373. Richards, J. E., Reynolds, G. D., & Courage, M. L. (2010). The neural bases of infant attention. Current Directions in Psychological Science, 19, 41–46. Robinson, A. K., Laning, J., Reinhard, J., & Mattingley, J. B. (2016). Odours reduce the magnitude of object substitution masking for matching visual targets in females. Attention, Perception, & Psychophysics, 78, 1702–1711. Rodieck, R. W. (1998). The first steps in seeing. Sunderland, MA: Sinauer. Rodrigues, J. F., da Silva Andrade, R., Bastos, S. C., Coelho, S. B., & Pinheiro, A. C. M. (2016). Miracle fruit: An alternative sugar substitute in sour beverages. Appetite, 107, 645–653. Rolls, E. T. (2006). Brain mechanisms underlying flavour and appetite. Philosophical Transactions of the Royal Society of London, Series B: Biological Sciences, 361, 1123–1136. Rolls, E. T., & Tovee, M. J. (1995). Sparseness of the neuronal representation of stimuli in the primate temporal visual cortex. Journal of Neurophysiology, 73, 713–726. Rosch, E. H. (1973). Natural categories. Cognitive Psychology, 4, 328–350. Sensation and Perception488 Rossi, E. A., & Roorda, A. (2010). The relationship between visual resolution and cone spacing in the human fovea. Nature Neuroscience, 13(2), 156. Rothbaum, B. O., Hodges, L., Alarcon, R., Ready, D., Shahar, F., Graap, K., . . . Baltzell, D. (1999). Virtual reality exposure therapy for PTSD Vietnam veterans: A case study. Journal of Traumatic Stress, 12(2), 263–271. Royster, J. D., Royster, L. H., & Killion, M. C. (1991). Sound exposures and hearing thresholds of symphony orchestra musicians. Journal of the Acoustical Society of America, 89, 2793–2803. Ruggeri, M., Major, J. C., Jr., McKeown, C., Knighton, R. W., Puliafito, C. A., & Jiao, S. (2010). Retinal structure of birds of prey revealed by ultra-high resolution spectral- domain optical coherence tomography. Investigative Ophthalmology & Visual Science, 51, 5789–5795. Rupp, M. A., McConnell, D. S., & Smither, J. A. (2016). Examining the relationship between action video game experience and performance in a distracted driving task. Current Psychology, 35(4), 527–539. Rushton, W. A. (1965). Visual adaptation. Proceedings of the Royal Society of London, Series B: Biological Sciences, 162, 20–46. Rutstein, R. P., & Daum, K. M. (1998). Anomalies of binocular vision: Diagnosis & management. St. Louis, MO: Mosby. Sabine, W. C. (1922). Collected papers on acoustics. Cambridge, MA: Harvard University Press. Sacks, O. (1985). The man who mistook his wife for a hat and other clinical tales. New York, NY: HarperPerennial. Sacks, O. (1998). The island of the colorblind. New York, NY: Vintage Books. Sacks, O. (2007). Musicophilia: Tales of music and the brain. New York, NY: Knopf. Sakurai, K., Kurita, T., Takeda, Y., Shiraishi, H., & Kusumi, I. (2013). Akinetopsia as epileptic seizure. Epilepsy & Behavior Case Reports, 1, 74–76. Salomão, S. R., Ejzenbaum, F., Berezovsky, A., Sacai, P. Y., & Pereira, J. M. (2008). Age norms for monocular grating acuity measured by sweep-VEP in the first three years of age. Arquivos Brasileiros de Oftalmologia, 71(4), 475–479. Salvi, S. M., Akhtar, S., & Currie, Z. (2006). Ageing changes in the eye. Postgraduate Medical Journal, 82(971), 581–587. Sandberg, K., Bahrami, B., Kanai, R., Barnes, G. R., Overgaard, M., & Rees, G. (2013). Early visual responses predict conscious face perception within and between subjects during binocular rivalry. Journal of Cognitive Neuroscience, 25, 969–985. Schachar, R. A. (2006). The mechanism of accommodation and presbyopia. International Ophthalmology Clinics, 46, 39–61. Schacter, D. L. (2001). Forgotten ideas, neglected pioneers: Richard Semon and the story of memory. Philadelphia, PA: Psychology Press. Schädler, M. R., Warzybok, A., Hochmuth, S., & Kollmeier, B. (2015). Matrix sentence intelligibility prediction using an automatic speech recognition system. International Journal of Audiology, 54(Suppl. 2), 100–107. Schenk, T., Ellison, A., Rice, N., & Milner, A. D. (2005). The role of V5/MT+ in the control of catching movements: An rTMS study. Neuropsychologia, 43, 189–198. Schenk, T., Mai, N., Ditterich, J., & Zihl, J. (2000). Can a motion-blind patient reach for moving objects? European Journal of Neuroscience, 12, 3351–3360. Schiells, R. A., & Falk, G. (1987). Joro spider venom: Glutamate agonist and antagonist on the rod retina of the dogfish. Neuroscience Letters, 77, 221–225. Schink, T., Kreutz, G., Busch, V., Pigeot, I., & Ahrens, W. (2014). Incidence and relative risk of hearing disorders in professional musicians. Occupational and Environmental Medicine, 71(7), 472–476. Schmidt, J. H. (2011). Sound exposure of symphony orchestra musicians. Annals of Occupational Hygiene, 55(8), 893–905. Schoenlein, R. W., Peteanu, L. A., Mathies, R. A., & Shank, C. V. (1991). The first step in vision: Femtosecond isomerization of rhodopsin. Science, 254, 412–415. Schultz, D. P., & Schultz, S. E. (1992). A history of modern psychology (5th ed.). New York, NY: Harcourt Brace. Schwartz, B. L. (2017). Memory: Foundations and applications (3rd ed.). Thousand Oaks, CA: Sage. Schwartz, D. A., & Purves, D. (2004). Pitch is determined by naturally occurring periodic sounds. Hearing Research, 194, 31–46. Schweizer, T., Kan, K., Hung, Y., Tam, F., Naglie, G., & Graham, S. (2013). Brain activity during driving with distraction: An immersive fMRI study. Frontiers in Human Neuroscience, 7(53), 1–11. Scott, M. A., van der Kamp, J., Savelsbergh, G. J., Oudejans, R.R.D., & Davids, K. (2004). Object rotation effects on the timing of a hitting action. Research Quarterly for Exercise and Sport, 75, 130–137. Seger, C. A., Spiering, B. J., Sares, A. G., Quraini, S. I., Alpeter, C., David, J., & Thaut, M. H. (2013). Corticostriatal contributions to musical expectancy perception. Journal of Cognitive Neuroscience, 25, 1062–1077. Setliff, A. E., & Courage, M. L. (2011). Background television and infants’ allocation of their attention during toy play. Infancy, 16, 611–639. Shepard, R. N. (1964). Circularity in judgements of relative pitch. Journal of the Acoustical Society of America, 36, 2346–2353. References 489 Shepard, R. N. (1982). Geometrical approximations to the structure of musical pitch. Psychological Review, 89, 305–333. Silverstein, L. D., & Merrifield, R. M. (1985). The development and evaluation of color systems for airborne applications (DOT/FAA Technical Report No. DOT/FAA/PM-85-19). Washington, DC: U.S. Department of Transportation. Simmons, J. A. (2012). Bats use a neuronally implemented computational acoustic model to form sonar images. Current Opinion in Neurobiology, 22, 311–319. Simner, J., Mulvenna, C., Sagiv, N., Tsakanikos, E., Witherby, S. A., Fraser, C., . . . Ward, J. (2006). Synaesthesia: The prevalence of atypical cross-modal experiences. Perception, 35, 1024–1033. Simons, D. J., & Chabris, C. F. (1999). Gorillas in our midst: Sustained inattentional blindness for dynamic events. Perception, 28, 1059–1074. Simons, D. J., & Levin, D. T. (1997). Change blindness. Trends in Cognitive Science, 1, 261–267. Singer, T., Seymour, B., O’Doherty, J., Kaube, H., Dolan, R. J., & Frith, C. D. (2004). Empathy for pain involves the affective but not sensory components of pain. Science, 303, 1157–1162. Slater, J., Azem, A., Nicol, T., Swedenborg, B., & Kraus, N. (2017). Variations on the theme of musical expertise: Cognitive and sensory processing in percussionists, vocalists and non-musicians. European Journal of Neuroscience, 45, 952–963. Smith, J. D., Redford, J. S., Washburn, D. A., & Taglialatela, L. A. (2005). Specific-token effects in screening tasks: Possible implications for aviation security. Journal of Experimental Psychology: Learning, Memory, and Cognition, 31, 1171–1185. Smith, V. C., & Pokorny, J. (2003). Color matching and color discrimination. In S. K. Shevell (Ed.), The science of color (pp. 103–148). New York, NY: Elsevier. Snyder, J. S., & Large, E. W. (2005). Gamma-band activity reflects the metric structure of rhythmic tone sequences. Cognitive Brain Research, 24, 117–126. Snyder, L. H., Batista, A. P., & Anderson, R. A. (2000). Intention-related activity in the posterior parietal cortex: A review. Vision Research, 40, 1433–1441. Sokol, S. (1978). Measurement of infant visual acuity from pattern reversal evoked potentials. Vision Research, 18(1), 33–39. Sommer, M. A., & Wurtz, R. H. (2008). Brain circuits for the internal monitoring of movements. Annual Review of Neuroscience, 31, 317–338. Stavrinos, D., Jones, J. L., Garner, A. A., Griffin, R., Franklin, C. A., Ball, D., Welburn, S. C., Ball, K. K., Sisiopiku, V. P., & Fine, P. R. (2013). Impact of distracted driving on safety and traffic flow. Accident Analysis & Prevention, 61, 63–70. Stein, B. E., & Meredith, M. A. (1993). The merging of the senses. Cambridge, MA: MIT Press. Stein, L. J., Cowart, B. J., Epstein, A. N., Pilot, L. J., Laskin, C. R., & Beauchamp, G. K. (1996). Increased liking for salty foods in adolescents exposed during infancy to a chloride-deficient feeding formula. Appetite, 27, 65–77. Stevens, K. A., & Brookes, A. (1988). The concave cusp as a determiner of figure-ground. Perception, 17, 35–42. Stevens, S. S. (1956). The direct estimation of sensory magnitudes: Loudness. American Journal of Psychology, 69, 1–25. Stevens, S. S. (1957). On the psychophysical law. Psychological Review, 64, 153–181. Stevens, S. S. (1961). To honor Fechner and repeal his law. Science, 133, 80–86. Stevenson, R. J. (2011). Olfactory illusions: Where are they? Consciousness and Cognition, 20, 1887–1898. Stevenson, R. J., & Mahmut, M. K. (2013). Detecting olfactory rivalry. Consciousness and Cognition, 22, 504–516. Stone, E. M. (2007). Macular degeneration. Annual Review of Medicine, 58, 477–490. Strayer, D. L., Watson, J. M., & Drews, F. A. (2011). Cognitive distraction while multitasking in the automobile. In B. H. Ross (Ed.), The psychology of learning and motivation: Advances in research and theory (Vol. 54, pp. 29–58). San Diego, CA: Academic Press. Streri, A., Lhote, M., & Dutilleul, S. (2000). Haptic perception in newborns. Developmental Science, 3(3), 319–327. Stroop, J. R. (1935). Studies of interference in serial verbal reactions. Journal of Experimental Psychology, 18(6), 643–662. Sunami, K., Ishii, A., Takano, S., Yamamoto, H., Sakashita, T., Tanaka, M., . . .Yamane, H. (2013). Neural mechanisms of phonemic restoration for speech comprehension revealed by magnetoencephalography. Brain Research, 1537, 164–173. Susilo, T., & Duchaine, B. (2013). Advances in developmental prosopagnosia research. Current Opinion in Neurobiology, 23, 423–429. Swaminathan, S. K., Masse, N. Y., & Freedman, D. J. (2013). A comparison of lateral and medial intraparietal areas during a visual categorization task. Journal of Neuroscience, 33, 13157–13170. Swettenham, J., Remington, A., Laing, K., Fletcher, R., Coleman, M., & Gomez, J. C. (2013). Perception of pointing from biological motion point-light displays in typically developing children and children with autism spectrum disorder. Journal of Autism and Developmental Disorders, 43, 1437–1446. Sensation and Perception490 Swiller, J. (2007). The unheard: A memoir of deafness and Africa. New York, NY: Henry Holt. Swithers, S. E. (2016). Not-so-healthy sugar substitutes? Current Opinion in Behavioral Sciences, 9, 106–110. Szycik, G. R., Stadler, J., Tempelmann, C., & Münte, T. F. (2012). Examining the McGurk illusion using high-field 7 tesla functional MRI. Frontiers in Human Neuroscience, 6, 95. Tamber-Rosenau, B. J., Esterman, M., Chiu, Y.-C., & Yantis, S. (2011). Cortical mechanism of cognitive control for shifting attention in vision and working memory. Journal of Cognitive Neuroscience, 23, 2905–2919. Tamietto, M., Cauda, F., Corazzini, L. L., Savazzi, S., Marzi, C. A., Goebel, R., . . . de Gelder, B. (2010). Collicular vision guides non-conscious behavior. Journal of Cognitive Neuroscience, 22, 888–902. Tan, S.-L., Pfordresher, P., & Harre, R. (2010). Psychology of music: From sound to significance. New York, NY: Psychology Press. Tapia, E., & Breitmeyer, B. G. (2011). Visual consciousness revisited: Magnocellular and parvocellular contributions to conscious and nonconscious vision. Psychological Science, 22, 934–942. Teller, D. Y. (1983). Scotopic vision, color vision, and stereopsis in infants. Current Eye Research, 2, 199–210. Temperley, D. (2000). Meter and grouping in African music: A view from music theory. Ethnomusicology, 44, 65–96. Terenius, L. (2000). From opiate pharmacology to opioid peptide physiology. Upsala Journal of Medical Sciences, 105(1), 1–15. Thornton, B., Faires, A., Robbins, M., & Rollins, E. (2014). The mere presence of a cell phone may be distracting: Implications for attention and task performance. Social Psychology, 45(6), 479–488. Tomita, T. (1970). Electrical activity of vertebrate photoreceptors. Quarterly Review of Biophysiology, 3, 179–222. Tong, F., Meng, M., & Blake, R. (2006). Neural bases of binocular rivalry. Trends in Cognitive Science, 10, 502–511. Tong, F., Nakayama, K., Vaughan, J. T., & Kanwisher, N. (1998). Binocular rivalry and visual awareness in human extrastriate cortex. Neuron, 21, 753–759. Toppila, E., Koskinen, H., & Pykkö, I. (2011). Hearing loss among classical-orchestra musicians. Noise and Health, 13, 45–50. Tramo, M. J. (2001). Biology and music: Enhanced: Music of the hemispheres. Science, 291, 54–56. Tranel, D., & Welsh-Bohmer, K. A. (2012). Pervasive olfactory impairment after bilateral limbic system destruction. Journal of Clinical and Experimental Neuropsychology, 34, 117–125. Treisman, A., & Gelade, G. (1980). A feature-integration theory of attention. Cognitive Psychology, 12, 97–136. Tremblay, P., Dick, A. S., & Small, S. L. (2011). New insights into the neurobiology of language from functional brain imaging. In H. Duffau (Ed.), Brain mapping: From neural basis of cognition to surgical applications (pp. 131–144). New York, NY: Springer. Troje, N. F. (2013). What is biological motion? Definition, stimuli and paradigms. In M. D. Rutherford & V. A. Kuhlmeier (Eds.), Social perception: Detection and interpretation of animacy, agency, and intention (pp. 13–36). Cambridge, MA: MIT Press. Troje, N. F., & Westhoff, C. (2006). The inversion effect in biological motion perception: Evidence for a “life detector”? Current Biology, 16, 821–824. Troncoso, X. G., Macknik, S. L., & Martinez-Conde, S. (2008). Microsaccades counteract perceptual filling-in. Journal of Vision, 8, 151–159. Turner, R. S. (1977). Hermann von Helmholtz and the empiricist vision. Journal of the History of the Behavioral Sciences, 13, 48–58. Turner, R. S. (1993). Vision studies in Germany: Helmholtz versus Hering. Osiris, 8, 80–103. Vaso, A., Adahan, H. M., Gjika, A., Zahaj, S., Zhurda, T., Vyshka, G., & Devor, M. (2014). Peripheral nervous system origin of phantom limb pain. PAIN, 155(7), 1384–1391. Vasterling, J. J., Proctor, S. P., Amoroso, P., Kane, R., Heeren, T., & White, R. F. (2006). Neuropsychological outcomes of army personnel following deployment to the Iraq war. JAMA, 206, 519–529. Venrooij, J., Mulder, M., Abbink, D. A., van Paassen, M. M., van der Helm, F. C. T., & Bülthoff, H. H. (2013). A new view on biodynamic feedthrough analysis: Unifying the effects on forces and positions. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, 43, 129–142. Verstraten, F. A. J. (1996). On the ancient history of the direction of the motion aftereffect. Perception, 25, 1177–1187. Võ, M. L., & Henderson, J. M. (2009). Does gravity matter? Effects of semantic and syntactic inconsistencies on the allocation of attention during scene perception. Journal of Vision, 9, 1–15. Võ, M. L., & Henderson, J. M. (2011). Object-scene inconsistencies do not capture gaze: Evidence from the flash-preview moving-window paradigm. Attention, Perception, & Psychophysics, 73, 1742–1753. von Arx, S. W., Müri, R. M., Heinemann, D., Hess, C. W., & Nyffeler, T. (2010). Anosognosia for cerebral achromatopsia: A longitudinal case study. Neuropsychologia, 48, 970–977. References 491 von Békésy, G. (1960). Vibratory pattern of the basilar membrane. In E. G. Wever (Trans. & Ed.), Experiments in hearing (pp. 404–429). New York, NY: McGraw-Hill. von der Heydt, R., Peterhans, E., & Baumgartner, G. (1984). Illusory contours and cortical neuron responses. Science, 244, 1260–1262. Voss, P., & Zatorre, R. J. (2012). Organization and reorganization of sensory-deprived cortex. Current Biology, 22, R168–R173. Wade, A., Augath, M., Logothetis, N., & Wandell, B. (2008). fMRI measurements of color in macaque and human. Journal of Vision, 8, 1–19. Wade, N. J. (1996). Descriptions of visual phenomena from Aristotle to Wheatstone. Perception, 25, 1137–1175. Wade, N. J., & Brožek, J. (2001). Purkinje’s vision: The dawning of neuroscience. Mahwah, NJ: Erlbaum. Wald, G. (1968). The molecular basis of visual excitation. Nature, 219, 800–807. Walls, G. L. (1963). The vertebrate eye and its adaptive radiation. Bloomfield Hills, MI: Cranbrook Institute of Science. Walsh, E., Long, C., & Haggard, P. (2015). Voluntary control of a phantom limb. Neuropsychologia, 75, 341–348. Wang, A., Duncan, S. E., & Dietrich, A. M. (2016). Effect of iron on taste perception and emotional response of sweetened beverage under different water conditions. Food Quality and Preference, 54, 58–66. Wang, C., & Yao, H. (2011). Sensitivity of V1 neurons to direction of spectral motion. Cerebral Cortex, 21, 964–973. Wang, Q., Schoenlein, R. W., Peteanu, L. A., Mathies, R. A., & Shank, C. V. (1994). Vibrationally coherent photochemistry in the femtosecond primary event of vision. Science, 266, 422–424. Wang, Y.-U., Chen, S.-A., & Lin, C.-T. (2014). An EEG-based brain–computer interface for dual task driving detection. Neurocomputing, 129, 85–93. Ward, L., Wright, E., & McMahon, S. B. (1996). A comparison of the effects of noxious and innocuous counterstimuli on experimentally induced itch and pain. Pain, 64, 129–138. Warren, R. M. (1970). Perceptual restoration of missing speech sounds. Science, 167, 392–393. Warren, W. H., Kay, B. A., Zosh, W. D., Duchon, A. P., & Sahuc, S. (2001). Optic flow is used to control human walking. Nature Neuroscience, 4, 213–216. Watanabe, M., & Rodieck, R. W. (1989). Parasol and midget ganglion cells in the primate retina. Journal of Comparative Neurology, 289, 434–454. Weeks, S. R., Anderson-Barnes, V. C., & Tsao, J. (2010). Phantom limb pain: Theories and therapies. Neurologist, 16, 277–286. Weigelt, S., Singer, W., & Kohler, A. (2013). Feature-based attention affects direction-selective fMRI adaptation in hMT+. Cerebral Cortex, 23, 2169–2178. Weiskrantz, L. (1986). Blindsight: A case study and implications. Oxford, UK: Clarendon. Weiskrantz, L. (1996). Blindsight revisited. Current Opinion in Neurobiology, 6, 215–220. Weiskrantz, L. (2009). Blindsight: A case study spanning 35 years and new developments. Oxford, UK: Oxford University Press. Weiskrantz, L., Warrington, E. K., Sanders, M. D., & Marshall, J. (1974). Visual capacity in the hemianopic field following a restricted occipital ablation. Brain, 97, 709–728. Weitzmann, K. (1970). Illustrations in roll and codex. Princeton, NJ: Princeton University Press. Werker, J. (2012). Perceptual foundations of bilingual acquisition in infancy. Annals of the New York Academy of Sciences: The Year in Cognitive Neuroscience, 1251, 50–61. Werker, J. F., & Gervain, J. (2013). Language acquisition: Perceptual foundations in infancy. In P. Zelazo (Ed.), The Oxford handbook of developmental psychology (pp. 909–925). Oxford, UK: Oxford University Press. Wert, K. J., Davis, R. J., Sancho-Pelluz, J., Nishina, P. M., & Tsang, S. H. (2013). Gene therapy provides long- term visual function in a pre-clinical model of retinitis pigmentosa. Human Molecular Genetics, 22, 558–567. Wertheimer, M. (1938). Laws of organization in perceptual forms. In W. Ellis (Ed.), A source book of gestalt psychology (pp. 71–88). London: Routledge & Kegan Paul. (Original work published 1923) Westhoff, C., & Troje, N. F. (2007). Kinematic cues for person identification from biological motion. Perception and Psychophysics, 69, 241–253. Wever, E. G., & Bray, C. W. (1937). The perception of low tones and the resonance-volley theory. Journal of Psychology, 3, 101–114. Wheatley, T., Weinberg, A., Looser, C. E., Moran, T., & Hajcak, G. (2011). Mind perception: Real but not artificial faces sustain neural activity beyond the N170/VPP. PLoS ONE, 6, e17960. Wiesel, T. N., & Hubel, D. H. (1963). Single cell responses in striate cortex of kittens deprived of vision in one eye. Journal of Neurophysiology, 26, 1003–1017. Wiesel, T. N., & Hubel, D. H. (1965). Comparison of the effects of unilateral and bilateral eye closure on cortical unit responses in kittens. Journal of Neurophysiology, 28, 1029–1040. Wilbiks, J. M. P., Vuvan, D. T., Girard, P.-Y., Peretz, I., & Russo, F. A. (2016). Effects of vocal training in a musicophile with congenital amusia. Neurocase, 22, 526–537. Sensation and Perception492 Williams, S. J., & Purves, D. (2001). Neuroscience. Sunderland, MA: Sinauer Associates. Wolfe, J. M. (2012). Saved by a log: How do humans perform hybrid visual and memory search? Psychological Science, 23, 698–703. Wong, A. O., & Wong, L. L. (2004). Tone perception of Cantonese-speaking prelingually hearing-impaired children with cochlear implants. Otolaryngology—Head and Neck Surgery, 130(6), 751–758. Wong-Riley, M. T. (1979). Changes in the visual system of monocularly sutured or enucleated cats demonstrable with cytochrome oxidase histochemistry. Brain Research, 171, 11–28. Woods, C. B., & Krantz, J. H. (2001). Sensation and perception: A window into the brain and mind. In J. S. Halonen & S. F. Davis (Eds.), The many faces of psychological research in the 21st century (Chap. 5). Retrieved from http://teachpsych.org/Resources/ Documents/ebooks/faces2001 Xiao, M., & Hendrickson, A. (2000). Spatial and temporal expression of short, long/medium, or both opsins in human fetal cones. Journal of Comparative Neurology, 425(4), 545–559. Xie, W.-J., Yu, H.-Q., Wang, Y.-H., Liu, Q., & Meng, H.-M. (2016). A case report of eight and a half syndrome. Neurological Sciences, 37, 821–822. Xu, X., Bonds, A. B., & Casagrande, V. A. (2002). Modeling receptive-field structure of koniocellular, magnocellular, and parvocellular LGN cells in the owl monkey (Aotus trivigatus). Visual Neuroscience, 19, 703–711. Xue, J., Peng, G., Yang, J. S., Ding, Q., & Cheng, J. (2013). Predictive factors of brain metastasis in patients with breast cancer. Medical Oncology, 30, 337. Yabuta, N. H., & Callaway, E. M. (1998). Functional streams and local connections of layer 4C neurons in primary visual cortex of the macaque monkey. Journal of Neuroscience, 11, 1352–1360. Yang, H., & Xu-Friedman, M. A. (2013). Stochastic properties of neurotransmitter release expand the dynamic range of synapses. Journal of Neuroscience, 33, 14406–14416. Yasuaki, T., & Seno, T. (2015). Roles of size, position, and speed of stimulus in vection with stimuli projected on a ground surface. Aerospace Medicine and Human Performance, 86, 794–802. Yasuo, T., Kusuhara, Y., Yasumatsu, K., & Ninomiya, Y. (2008). Multiple receptor systems for glutamate detection in the taste organ. Biological & Pharmaceutical Bulletin, 31, 1833–1837. Yin, T. C., & Chan, J. C. (1990). Interaural time sensitivity in medial superior olive of cat. Journal of Neurophysiology, 65, 465–488. Yoshida, R., & Ninomiya, Y. (2010). New insights into the signal transmission from taste cells to gustatory nerve fibers. International Review of Cell and Molecular Biology, 279, 101–134. Yost, W. A. (2007). Fundamentals of hearing: An introduction (5th ed.). New York, NY: Academic Press. Zakis, J. A., Fulton, B., & Steele, B. R. (2012). Preferred delay and phase-frequency response of open- canal hearing aids with music at low insertion gain. International Journal of Audiology, 51, 906–913. Zamm, A., Schlaug, G., & Eagleman, D. M. (2013). Pathways to seeing music: Enhanced structural connectivity in colored-music synesthesia. Neuroimage, 74, 359–366. Zanos, T. P., Mineault, P. J., Guitton, D., & Pack, C. C. (2016). Mechanisms of saccadic suppression in primate cortical area V4. Journal of Neuroscience, 36(35), 9227–9239. Zeki, S. (1993). A vision of the brain. Oxford, UK: Blackwell. Zheng, J., Zhang, B., Bi, H., Maruko, I., Watanabe, I., Nakasata, C., . . . Chino, Y. M. (2007). Development of temporal response properties and contrast sensitivity of V1 and V2 neurons in macaque monkeys. Journal of Neurophysiology, 97, 3905–3916. Zhou, W., & Chen, D. (2009). Binaral rivalry between the nostrils and in the cortex. Current Biology, 19, 1561–1565. Zihl, J., von Cramon, J. D., & Mai, N. (1983). Selective disturbance of movement vision after bilateral brain damage. Brain, 106, 313–340. Zivony, A., & Lamy, D. (2014). Attentional engagement is not sufficient to prevent spatial capture. Attention, Perception, & Psychophysics, 76, 19–31. Zokoll, M. A., Wagener, K. C., Brand, T., Buschermöhle, M., & Kollmeier, B. (2012). Internationally comparable screening tests for listening in noise in several European languages: The German digit triplet test as an optimization prototype. International Journal of Audiology, 51, 697–707. Zucco, G. M., Aiello, L., Turuani, L., & Koster, E. (2012). Odor-evoked autobiographical memories: Age and gender differences along the life span. Chemical Senses, 37(2), 179–189. doi:10.1093/chemse/bjr089 Author Index 493 AUTHOR INDEX Abbink, D. A., 16 Abdelaziz, A. H., 365 Abel, G. L., 421 Abramov, I., 164 Adahan, H. M., 428 Adam, A., 180 Adelson, E. H., 177 Agostini, T., 178 Ahrens, W., 394 Aiello, L., 45 Akhtar, S., 62 Aladdin, Y., 117 Alarcon, R., 221 Albouy, G., 421 Alexander, E., 342 Alkaladi, A., 83 Alpern, M., 171, 174 Alpeter, C., 384 Amoros, P., 18 Andersen, R. A., 245 Anderson, B. A., 265 Anderson, R. A., 245 Anderson, S., 363 Anderson-Barnes, V. C., 427 Anstis, S., 248 Ardiles, P., 22 Armony, J. L., 277 Arnell, K. M., 257 Arnold, K., 152 Arnoldussen, A., 86 Arroyo, J. G., 80 Atkinson, J., 72 Attems, J., 454 Augath, M., 175 Axel, R., 439–440 Azem, A., 385 Babadi, B., 97 Bacon, S. P., 330 Bahrami, B., 281 Baldwin, G. T., 417 Balkwill, L.-L., 391 Ball, D., 282 Ball, K. K., 282 Ball, S. L., 452 Baltzell, D., 221 Banks, M. S., 72 Bannister, J., 27 Barden, K., 354 Barlow, H. B., 138, 210, 232 Barnes, G. R., 281 Barry, M. P., 85 Bartoshuk, L. M., 449, 451, 452 Bastos, S. C., 456–457 Batista, A. P., 245 Battelli, L., 242 Baumgartner, G., 135 Baures, R., 22 Baxter, L. C., 415 Beauchamp, G. K., 453 Beck, D. M., 270 Beckers, G., 239 Bekrater-Bodmann, R., 428 Belin, P., 361 Bennett, C. M., 19 Bennett, P. J., 72 Bensmaia, S., 420 Beran, M. J., 56 Beranek, L. L., 336 Berezovsky, A., 72 Bernotas, M., 199 Bernstein, I. L., 453 Berson, E. L., 81 Bestelmeyer, P. E. G., 361 Bhalla, M., 21 Bi, H., 211 Biederman, I., 136 Birch, E. E., 170, 210 Bisiach, E., 276 Bisley, J. W., 106 Blackwell, H. R., 36 Blake, R., 82, 241, 281 Blakemore, C., 210 Bleyenheuft, Y., 422 Bohn, K. M., 334 Bonafede, B., 218–219 Bonato, F., 178 Bonds, A. B., 97 Boot, W. R., 146–147 Boring, E. G., 211–212 Born, R. T., 110 Bornstein, M. H., 168 Bosten, J. M., 152, 174 Botella, C., 221 Bouchard, S., 221 Boucsein, W., 250 Bowmaker, J. K., 66 Boyer, A., 215 Boyle, J., 445 Brackett, D. D., 215 Braddick, O., 72 Bradley, D. C., 110 Bramanti, P., 454 Brand, T., 48 Brang, D., 386 Bray, C. W., 305 Bregman, A. S., 327, 328, 342 Breitmeyer, B. G., 44, 232 Brendel, E., 22 Breslin, P. A. S., 452 Bressler, N. M., 80 Brett, J. A., 21 Bridgeman, B., 235 Brincat, S., 137 Britten, K. H., 237, 238 Brookes, A., 132 Brown, A. M., 169 Browne, J. V., 330 Brozek, J., 69 Bruce, C., 109 Buck, L. B., 439–440 Bucy, P. C., 137–138 Buetti, S., 281 Buletic, Z., 452 Bülthoff, H. H., 16 Bundesen, C., 231 Busch, N. A., 270 Busch, V., 394 Buschermöhl, M., 48 Bushnell, M. C., 416 Busigny, T., 139 Butler, J., 233 Buzás, P., 152 Cahn, R., 384 Cain, W. S., 444 Cain Miller, C., 125 Calkins, D., 106 Callaway, E. M., 98 Caminiti, F., 454 Cantril, H., 4 Carey, D. P., 109 Carrier, B., 416 Carroll, J., 171 Cartwright, M., 372 Carylon, R. P., 342 Casagrande, V. A., 97 Casti, A., 97 Cataliotti, J., 178 Cauda, F., 115, 281 Cave, K. R., 146 Cavina-Pratesi, C., 110 Centers for Disease Control and Prevention, 306 Chabris, C. F., 263 Champoux, F., 385 Chan, B. L., 428 Chan, J. C., 323 Chandrashekar, J., 450 Chang, E. F., 361–362, 367 Charrow, A. P., 428 Chasin, M., 363, 395–396 Chen, D., 446 Chen, D. W., 98 Chen, M.-C., 277 Chen, S.-A., 283 Chen, X., 20 Cheng, H., 210–211 Cheng, J., 42 Cherrier, M. M., 140 Cherry, E. C., 342 Chino, Y. M., 210–211 Chiu, Y.-C., 273, 274 (figure) Chow, A., 85 Chow, V., 85 Chrastil, E. R., 244 Chun, M. M., 138, 257 Ciurleo, R., 454 Claggett, B., 20 Clark, J., 270 Coben, J. H., 282 Coe, S. B., 456–457 Colby, C. L., 245 Cole, B. L., 180 Cole, J., 428 Collignon, O., 421 Sensation and Perception494 Colombo, J., 278 Compton, W. M., 417 Conard, N. J., 372 Connor, C. E., 109, 137 Cool, S. J., 82 Corazzini, L. L., 115, 281 Corbit, J. D., 350 Cornsweet, T. N., 71 Côté, S., 221 Coughlin, B. C., 452 Courage, M. L., 278, 279, 285 Cowart, B. J., 453 Craig, A. D., 415 Craig, J. C., 420 Crawford, M. L. J., 82 Crawley, A. P., 416 Crone, N. E., 361–362, 367 Croy, I., 455 Crystal, S. R., 453 Cui, H., 245 Culham, J. C., 110, 246 Curcio, C. A., 64 Currie, Z., 62 Curtiss, S., 358 Da Cruz, L., 85 Dagnelie, G., 85 Dai, P., 366 Damasio, A., 384 Damasio, H., 384 Da Silva Andrade, R., 456–457 Daum, K. M., 61 David, D., 221 David, J., 384 David, S. V., 361–362, 367 Davids, K., 219 Davis, K. D., 416 Davis, R. J., 81 Daw, N., 113 DeCasper, A. J., 330 De Courten, C., 113 Deeb, S. S., 152, 174 De Gelder, B., 113–114, 115, 281 Dekle, D. J., 352 Delk, J. L., 20 Delogu, F., 456 DeLucia, P. R., 22 Delwiche, J. F., 452 DeMarco, P. J., Jr., 85 Dennett, D. C., 111 Desimone, R., 109, 275, 284 Deutsch, D., 392–394 De Valois, R. L., 164, 166 Devor, M., 428 Devos, M., 444 Dewald, A. D., 257 De Wijk, R. A., 444 DeWitt, L. A., 383 DeYoe, E. A., 107 Dick, A. S., 18 Diehl, R. L., 355 Diers, M., 428 Dijkstra, P. U., 427 Dilks, D., 138 Di Lollo, V., 44 Ding, Q., 42 Ditterich, J., 239 Djordjevic, J., 445 Dobson, V., 72 Doerschner, K., 16 Dolan, R. J., 140, 416 Dominguez-Borràs, J., 277 Donnelly, N., 146 Dormal, G., 421 Dorn, J. D., 85 Dostrovsky, J. O., 416 Doumas, L. A. A., 257 Downing, P. E., 140 Drew, T., 264 Drews, F. A., 256 Driver, J., 140, 261–262, 277, 428 Dryja, T. P., 81 Duarte, N., 221 Duchaine, B. C., 110, 139–140 Duchon, A. P., 244 Duffy, V. B., 451, 452 Duhamel, J.-R., 245 Duncan, G. H., 416 Dutilleul, S., 422 Eagleman, D. M., 386 Economou, E., 178 Eddy, R., 64 Eggenberger, E. R., 116 Egly, R., 261–262 Egorova, N., 20 Eimas, P. D., 350 Ejzenbaum, F., 72 Ellison, A., 110 Elzinga, A., 427 Engen, T., 446 Enns, J. T., 44 Epstein, A. N., 453 Epstein, R., 140 Erberich, S., 385 Esterman, M., 273, 274 (figure) Fain, G. L., 66 Faires, A., 283 Falk, G., 67 Farina, F. R., 386 Fark, T., 454 Fattori, P., 235 Fechner, G. T., 13 Fedderson, W. E., 323 Fernandez, R., 456 Feynman, R. P., 55 Fields, R. D., 426 Fifer, W. P., 330 Fillenbaum, S., 20 Findling, D., 125, 126 Fine, P. R., 282 Firestone, C., 20–21 Fisher, M. F. K., 447 Flanagan, J. R., 246 Flinker, A., 361–362, 367 Flor, H., 428 Foell, J., 428 Foldvari, M., 98 Foltys, H., 385 Foster, D. H., 176 Fowler, C. A., 352, 355, 356 Frank, G. K. W., 456 Franklin, C. A., 282 Fraser, A., 247 Fraser, C., 386 Frasnelli, J., 454 Frazier, L. D., 455 Freedman, D. J., 246 Freeman, S., 20 Fried, I., 110, 141, 142 Frith, C. D., 416 Froehlich, J. C., 417 Fromkin, V., 358 Fudge, J. L., 456 Fujita, N., 419 Fulton, B., 395 Furl, N., 140 Galantucci, B., 355 Galletti, C., 235 Gallivan, J. P., 246 Gamito, P., 221 Garcia-Palacios, A., 221 Garneau, N. L., 452 Garner, A. A., 282 Garrido, L., 140 Gartshteyn, Y., 237 Gatzia, D. E., 20 Geertzen, J. H. B., 427 Gegenfurtner, K. R., 20 Gelade, G., 267 Gelbard-Sagiv, H., 142 Gelfand, S. A., 32 Gerber, J., 454 Gerstmann, J., 421 Gervain, J., 357 Gibson, E. J., 15 Gibson, J. J., 5, 194, 198, 243, 356 Gilad, Y., 440 Gilbert, S. F., 62 Gilchrist, A., 178 Girard, P.-Y., 387 Gjika, A., 428 Godfrey, P. A., 440 Godwin, H. J., 146–147 Goebel, R., 115, 281 Goldberg, M. E., 245 Goldberg, S., 116–117 Goldinger, S. D., 146–147 Goldwater, S., 365 Golomb, J. D., 257 Goodale, M. A., 109 Gorman, J. C., 22 Gottfried, J. A., 442 Graap, K., 221 Graham, C. H., 174 Graham, S., 283 Grant, P., 86 Granville, W. C., 169 Graven, S. N., 330 Green, B. G., 418 Green, D. M., 38 Green, E., 456 Greenberg, R. J., 85 Author Index 495 Greenwood, V., 152 Gregory, R. L., 214 Griffin, R., 282 Grill-Spector, K., 138 Grimault, N., 330 Groen, E. L., 250 Gross, C. G., 109 Grossman, E. D., 242 Grueschow, M., 442 Grüsser, O.-J., 11 Guitton, D., 259 Habibi, A., 384 Hachem, S., 456 Haehner, A., 455 Haggard, P., 427 Hahnel, U. J. J., 22 Hajcak, G., 144 Hakyemez, H., 342 Hanif, W., 44 Hansen, T., 20 Harel, M., 142 Harley, T. A., 357, 358 Harre, R., 381–382, 390 Harris, A., 138 Hartline, H. K., 76, 84 Hartong, D. T., 81 Haslwanter, T., 199 Håstad, O., 56 Hastorf, A. H., 4 Hatta, S., 210–211 Hawken, M. J., 167 Hayes, J. E., 452 Haynes, J. D., 442 He, J., 319 Hecht, H., 22 Hecht, S., 67 Hedges, J. H., 237 Heeren, T., 18 Heilman, K. M., 428 Hein, A., 112 Heinemann, D., 175 Heinze, H. J., 19 Held, R., 112 Heller, M. A., 215 Helman, S., 22 Henderson, J. M., 266 Hendrickson, A. E., 64, 168 Hendry, S. H. C., 95 Hershenson, M., 250 Hervais-Adelman, A., 281 Herz, R., 435–436, 444–445, 453–455 Herz and von clef, 446 Herz eliassen beland and souza, 443 Hess, C. W., 175 Hickok, G., 361 Highnote, S., 68, 174 Hirstein, W., 427 Hochmuth, S., 365 Hockley, N. S., 395–396 Hodges, L., 221 Hoefle, F. B., 218–219 Hof, P. R., 411 Hofeldt, A. J., 218–219 Hofer, H., 171 Hoffman, P. M., 325 Hogle, R., 86 Hogness, D., 64 Hollins, M., 420 Holt, L. L., 355, 356 Holway, A. H., 211–212 Hoon, M. A., 450 Hopf, J. M., 19 Horridge, A., 56 Horswill, M., 22 Houpt, J. W., 146–147 Hout, M. C., 146–147 Howard, I. P., 205 Howard, J. D., 442, 443 Howard, R., 428 Hsia, Y., 174 Hsiang, S., 22 Hsiao, S., 414–415 Hsiao, S. S., 420 Hu, L., 385 Hua, L. H., 415 Huang, Q., 364 Huang, Y.-T., 277 Hubbard, E. M., 386 Hubel, D. H., 17, 91, 103–104, 105, 112–113, 209 Huddas, C., 456 Hughes, H. C., 260–261, 334, 426 Humayun, M. S., 85 Hummel, T., 454 Hung, Y., 283 Huotilainen, M., 330 Hurvich, L. M., 165 Hutchinson, W. D., 416 Huttenlocher, P. R., 113 Hyde, K., 384, 386 Iannetti, G. D., 385 Irons, J., 271 Isaacs, L. D., 219 Ishii, A., 354 Isono, K., 188 Jacobs, G., 64 Jacobs, G. H., 164 Jacquin-Cortois, S., 277 Jakobsen, L. S., 109 James, W., 278 Jameson, D., 165 Jameson, K. A., 68, 174 Jeffress, L. A., 323 Jellinger, K. A., 454 Jiang, Y., 140 Jiao, S., 65 Johansson, G., 241 Johnson, E. N., 167 Johnson, S. C., 415 Johnsrude, I. S., 342 Jolicoeur, P., 387 Jones, C. M., 417 Jones, F. N., 36 Jones, J. L., 282 Jones-Gotman, M., 445 Jönsson, F. U., 444 Jordan, G., 152, 174 Juang, B. H., 365 Julesz, B., 208–209 Jurafsky, D., 365 Kaas, A., 238 Kadohisa, M., 442 Kalina, R. E., 64 Kallus, K. W., 250 Kan, K., 283 Kanai, R., 281 Kandel, E. R., 101 Kandil, F. I., 244 Kane, R., 18 Kanizsa, G., 135 Kanwisher, N., 110, 138, 140, 281 Kaplan, E., 97 Kapoula, Z., 199 Kaptchuk, T. J., 20 Kau, S., 19 Kaube, H., 416 Kawahara, J. I., 272 Kay, B. A., 244 Kaye, W. H., 456 Keane, A. G., 21 Kennedy, H., 233 Kersten, D., 16 Kerzel, D., 281 Kessen, W., 168 Khateb, A., 116, 281 Killion, M. C., 395 Kirchhof, H., 61 Kirsch, I., 20 Kitahara, K., 174 Kitaoka, A., 247 Klatzky, R. L., 419 Klüver, H., 137–138 Knighton, R. W., 65 Knouf, N., 138 Kóbor, P., 152 Koch, C., 110, 141 Kochkin, S., 362 Kohler, A., 238 Kohn, A., 237 Kollmeier, B., 48, 365 Kolossa, D., 365 Kong, J., 20 Kooijman, C. M., 427 Koskinen, H., 395 Kossyfidis, C., 178 Koster, E., 45 Kovach, J. K., 10 Koyanagi, M., 188 Kramer, A. F., 146–147 Krantz, D. H., 174 Krantz, J. H., 72, 170 Krashen, S., 358 Kraus, N., 363, 385 Kreiman, G., 110, 141 Kreutz, G., 394 Krings, T., 385 Kristjánsson, A., 272, 283 Krumhansl, C. L., 390 Krupinski, E. A., 146 Kuffler, S. W., 17, 73–74, 76, 82 Kujalaa, T., 330 Kundel, H. L., 146 Sensation and Perception496 Kurita, T., 239 Kusumi, I., 239 Laffort, P., 444 LaMotte, R. H., 418 Lancet, D., 440 Land, E. H., 177 Landis, B. N., 455 Landry, S. P., 385 Langenheim, J., 298 Laning, J., 44–45 Lappe, M., 244 Large, E. W., 385 Larsen, A., 231 Larsen, J. T., 22 Laskin, C. R., 453 Latinus, M., 361 Laurent, P. A., 265 Lavie, N., 270 Lawless, H. T., 445 Lazeyras, F., 116, 281 Lederman, S. J., 419 Lee, G. B., 171 Leek, M. R., 364 Leh, S. E., 116 Lehmann, J., 454 Lénárd, L., 152 Lenoir, M., 219 Lepore, F., 421 Lettvin, J. Y., 232 Levin, D. T., 269 Lhote, M., 422 Li, W., 443 Li, X., 178 Liang, M., 385 Liberman, A. M., 355 Limb, C. J., 396 Lin, C.-T., 283 Lin, K., 277 Litola, A., 330 Liu, J., 138 Liu, Q., 117 Livadas, C., 250 Livingstone, M., 105, 179 Lodhia, L., 456 Logerstedt, M., 456 Logothetis, N., 175 Lokki, T., 335 Long, C., 427 Loomis, J. M., 419 Looser, C. E., 143–145, 144 Lopez, L., 384, 443 Lotto, A. J., 355, 356 Loui, P., 386 Lowenfield, B., 421 Lozano, A. M., 416 Luaté, J., 277 Lucchina, L. A., 452 Lund, T. E., 231 Luzzatti, C., 276 Lygeros, J., 250 Lynch, N. A., 250 Maaseidvaag, F., 171 MacDonald, T., 351–352 Mack, A., 263, 264 (figure) Mackey, A., 342 Macknik, S. L., 247, 248 Macko, K. A., 107–109 MacMahon, M. K. C., 346 Madsen, K. H., 231 Maertens, M., 135 Magee, A., 428 Mahmut, M. K., 446 Mai, N., 239 Majno, M., 384 Major, J. C., Jr., 65 Malach, R., 142 Malina, M., 372 Malnic, B., 440 Man, O., 440 Manning, C. D., 365 Maravita, A., 277 Marchi, F., 20 Marino, S., 454 Marr, D., 16 Marshall, J., 114 Martin, P. R., 152 Martinez-Conde, S., 247, 248 Maruko, I., 211 Marzi, C. A., 115, 281 Masland, R. H., 232 Mason, C., 101 Masse, N. Y., 246 Mathies, R. A., 67 Mattingley, J. B., 44–45 Mattingly, I. G., 355 Mattys, S. L., 354 Maturana, H. R., 232 Maurer, D., 357 Mayer, D. L., 72 Mazo, M., 391 Mazyn, L. I. N., 219 McCall, M. A., 452 McCarley, J. S., 146–147 McConnell, D. S., 283 McCulloch, W. S., 232 McDermott, J., 138 McGlone, F., 406 McGurk, H., 351–352 McKeegan, N., 85 McKeown, C., 65 McLean, D. A., 246 McLean, G. Y., 452 McMahon, S. B., 418 Melzack, R., 410, 415 Memmi, S., 241 Menashe, I., 440 Mendez, M. F., 140 Meng, H.-M., 117 Meng, M., 281 Menneer, T., 146 Mercier, C., 428 Merkel, C., 19 Merrifield, R. M., 159 Mesgarani, N., 361–362, 367 Meyer, G. E., 135 Meyer, R., 336, 391 Meza-Arroyo, M., 22 Micheyl, C., 330 Mignault-Goulet, G., 387 Mikulis, D. J., 416 Miller, F. G., 20 Miller, I. J., 451 Miller, L. R., 20 Miller, M. B., 19 Miller, S. S., 171 Milner, A. D., 109, 110 Mineault, P. J., 259 Misaka, T., 457 Mishkin, M., 107–109 Mitchell, K. J., 386 Moar, K., 72 Molis, M. R., 364 Møller, J., 308 Mollon, J. D., 152, 161, 174, 180 Monnier, P., 179 Montagne, G., 219 Moore, B. C. J., 396 Morais, D., 221 Moran, J., 275, 284 Moran, T., 144 Moreau, P., 387 Morgan, M. J., 180 Morrison, I., 406 Moscovitch, D., 139 Moscovitch, M., 139 Mountcastle, V. B., 421 Mouraux, A., 385 Movshon, J. A., 237, 238 Muayqil, T., 117 Muckli, L., 238 Muggleton, N., 270 Mukamel, R., 142 Mulder, M., 16 Mulvenna, C., 386 Münte, T. F., 352 Münzel, S. C., 372 Müri, R. M., 175 Murphy, C., 456 Murphy, G., 10 Näätänen, R., 330 Nagata, T., 188 Naglie, G., 283 Nakasata, C., 211 Nakayama, K., 139–140, 281 Nassi, J. J., 98 Nathans, J., 64 National Institutes of Health, 46 Nau, A., 86 Neisser, U., 263 Neitz, J., 64, 171 Neitz, M., 171 Neumeyer, C., 57, 152 Newen, A., 20 Newsome, W. T., 237–238 Nicol, T., 385 Nilsson, D.-E., 83 Ninomiya, Y., 450 Nishina, P. M., 81 Nodine, C. F., 146 Nooij, S. A. E., 250 Nuessle, T. M., 452 Nyffeler, T., 175 Author Index 497 Oberndorfer, T. A., 456 Ödeen, A., 56 O’Doherty, J., 416 Olausson, H., 406 Oliveira, J., 221 Oliveira, S., 221 Olkkonen, M., 20 Olsson, M. J., 444 Oltman, M., 435 Opris, D., 221 Orban, G. A., 233 O’Regan, J. K., 270 O’Shea, J., 277 Osher, J., 443 Otero-Millan, J., 247 Oudejans, R. R. D., 219 Overgaard, M., 281 Overy, K., 384 Pack, C. C., 259 Packo, K., 85 Palmer, C., 391 Paninski, L., 97 Parbery-Clark, A., 363 Paré, E. B., 237–238 Parker, A. J., 210 Parkinson, C., 241 Partanen, E., 330 Pascual-Leone, A., 242 Pasley, B., 361–362, 367 Pasquina, P. F., 428 Pasternak, T., 106 Pasupathy, A., 109, 137 Patel, A. D., 388–389 Patte, F., 444 Patynen, J., 335 Paulus, M. P., 456 Pavani, F., 428 Pegna, A. J., 116, 281 Pei, Y. C., 420 Penfield, W., 413 Peng, G., 42 Pereira, J. M., 72 Peretz, I., 384, 386, 387 Perlman, M., 390 Perry, B. N., 428 Peteanu, L. A., 67 Peterhans, E., 135 Peterson, M. A., 132 Petrides, M., 445 Petry, S., 135 Pettifer, S. R., 428 Pettigrew, J. D., 210 Petykó, Z., 152 Peyman, G., 85 Pfordresher, P., 381–382, 390 Phillips, C., 421 Piantanida, T., 64 Pigeot, I., 394 Pilot, L. J., 453 Pinckers, A. J. L. G., 169 Pinheiro, A. C. M., 456–457 Pinker, S., 347 Pintea, S., 221 Pirenne, M. H., 67 Pisella, L., 277 Pitts, W. H., 232 Plailly, J., 442 Plantinga, J., 381 Poeppel, D., 361 Pokorny, J., 173 Pollack, J., 85 Pollmann, S., 135 Posner, M., 260–262, 273–274 Price, D. D., 416 Proctor, R. W., 180 Proctor, S. P., 18 Proffitt, D. R., 21 Ptito, A., 116 Puliafito, C. A., 65 Purves, D., 296, 409 Pykkö, I., 395 Quiroga, R. Q., 110, 141, 245 Quraini, S. I., 384 Rabiner, L. R., 365 Radtke, N. D., 85 Rafal, R. D., 261–262 Rainville, P., 416 Ramachandran, V. S., 65, 386, 427 Ranjit, M., 22 Rasmussen, T., 413 Ratliff, F., 76, 84 Rauschecker, J. P., 321 Raymond, J. E., 257 Rayner, K., 146 Read, J. C. A., 104 Ready, D., 221 Reddy, L., 110, 141 Reed caselli farah, 421 Rees, G., 281 Reid, R. C., 95 Reinhard, J., 44–45 Remington, R., 271 Remus, D. R., 281 Renaud, P., 221 Rensink, R. A., 270 Revol, P., 277 Reynolds, G. D., 278, 285 Rice, N., 110 Richards, J. E., 278, 285 Rigler, D., 358 Rigler, M., 358 Robbins, M., 283 Robillard, G., 221 Robinson, A. K., 44–45 Robitaiile, N., 387 Roche, R. A. P., 386 Rock, I., 263, 264 (figure) Rodieck, R. W., 66, 73 Rodrigues, J. F., 456–457 Roebroeck, A., 238 Rogers, B., 248 Rogers, B. J., 205 Rollins, E., 283 Rolls, E. T., 450 Rolls and tovee, 138 Roorda, A., 70, 72 Rosa, P., 221 Rosch, E. H., 164 Rossetti, Y., 277 Rossi, E. A., 70, 72 Rossion, B., 139 Rothbart, M. K., 273 Rothbaum, B. O., 221 Rotter, A., 244 Rouault, J., 444 Roy, A. T., 396 Royster, J. D., 395 Royster, L. H., 395 Ruggeri, M., 65 Rupp, M. A., 283 Russo, F. A., 387 Rust, N. C., 237 Rutstein, R. P., 61 Ryba, N. J. P., 450 Sabine, W. C., 335 Sacai, P. Y., 72 Sacks, O., 123, 171, 276 Saeki, S., 188 Sagdullaev, B. T., 452 Sagiv, N., 386 Sahel, J. A., 85 Sahuc, S., 244 Sakashita, T., 354 Sakurai, K., 239 Salomão, S. R., 72 Salvagio, E., 132 Salvi, S. M., 62 Sambeth, A., 330 Samuel, A. G., 354, 383 Samuels, S. M., 250 Sancho-Pelluz, J., 81 Sandberg, K., 281 Sandel, T. T., 323 Sanders, M. D., 114 Santorico, S. A., 452 Saqqur, M., 117 Saraiva, T., 221 Sares, A. G., 384 Sato, H., 272 Savazzi, S., 115, 281 Savelsbergh, G. J. P., 22, 219 Schachar, R. A., 200 Schacter, D. L., 141 Schädler, M. R., 365 Schenk, T., 110, 239 Schiells, R. A., 67 Schink, T., 394 Schlaer, S., 67 Schlaug, G., 386 Schmidt, J. H., 395 Schoenfeld, M. A., 19 Schoenlein, R. W., 67 Scholl, B. J., 20–21 Schrater, P., 16 Schultz, D. P., 13 Schultz, S. E., 13 Schwartz, B. L., 20, 110, 127 Schwartz, D. A., 296 Schweizer, T., 283 Scott, M. A., 219 Scott, S. K., 321 Sensation and Perception498 Scozzafava, J., 117 Seger, C. A., 384 Seghier, M., 116, 281 Seiple, W., 86 Seno, T., 250 Setliff, A. E., 279, 285 Seymour, B., 416 Shadlen, M. N., 237 Shahar, F., 221 Shamma, S. A., 361–362, 367 Shank, C. V., 67 Shapiro, K. L., 257 Shapley, R., 167 Shepard, R. N., 375 (figure), 392 Shichida, Y., 188 Shiffrar, M., 241 Shiller, D. M., 385 Shimada, S. G., 418 Shiraishi, H., 239 Shows, T., 64 Shuman, M., 140 Silverstein, L. D., 72, 159 Simmons, A. N., 456 Simmons, J. A., 331 Simner, J., 386 Simons, D. J., 263, 269 Singer, T., 416 Singer, W., 238 Singhai, A., 110 Sinnett, S., 257 Sisiopiku, V. P., 282 Slater, J., 385 Sloan, K. R., 64 Sloan, M. M., 452 Small, S. L., 18 Smarsh, G. C., 334 Smith, E. L., III, 210–211 Smith, V. C., 173 Smither, J. A., 283 Smith redford washburn and taglialatela, 146 Smotherman, M. S., 334 Snyder, J. S., 385 Snyder, L. H., 245 Sokol, S., 72 Sommer, M. A., 235 Soon, I. Y., 366 Sparing, R., 385 Spence, C., 428 Spencer, L., 86 Spiering, B. J., 384 Stacy, R. L., 22 Stadler, J., 352 Stark, L., 235 Starr, A., 384 Stavrinos, D., 282 Steele, B. R., 395 Steffen, H., 215 Stein, L. J., 453 Steven, K., 456 Stevens, K. A., 132 Stevens, S. S., 34, 35 Stevenson, R. J., 444–446 St-Jacques, J., 221 Stone, E. M., 80 Stoppel, C. M., 19 Stranga, P. A., 85 Strayer, D. L., 256 Streri, A., 422 Streri lhote dutilleul, 422 Strigo, I. A., 415 Stroop, J. R., 258 Stroud, M., 146 Strumpf, H., 19 Sunami, K., 354 Susilo, T., 110 Swaminathan, S. K., 246 Swedenborg, B., 385 Swets, J., 38 Swiller, J., 306 Swithers, S. E., 455, 457 Szamosközi, S., 221 Szlyk, J., 86 Szycik, G. R., 352 Takano, S., 354 Takeda, Y., 239 Tam, F., 283 Tamber-Rosenau, B. J., 273, 274 (figure) Tamietto, M., 115, 281 Tan, S.-L., 381–382, 390 Tanaka, M., 354 Tang, J., 364 Tan pfordresher harre, 373 Tapia, E., 232 Tasker, R. R., 416 Taylor, S. J., 416 Teas, D. C., 323 Telkes, I., 152 Teller, D. Y., 169, 210 Tempelmann, C., 352 Temperley, D., 390 Terakita, A., 188 Terenius, L., 417 Thaut, M. H., 384 Thompson, J. M., 356 Thompson, W. F., 391 Thonnard, J. L., 422 Thornton, B., 283 Thron, A., 385 Tomita, T., 67 Tong, F., 281 Töpper, R., 385 Toppila, E., 395 Trainor, L. J., 381 Tramo, M. J., 385 Tranel, D., 454 Trang, H. P., 342 Trattler, W., 116–117 Treisman, A., 267 Tremblay, P., 18 Troje, N. F., 241, 242 Troncoso, X. G., 248 Tropper, K., 250 Tsai, P.-L., 277 Tsakanikos, E., 386 Tsang, S. H., 81 Tsao, J., 427 Tsao, J. W., 428 Tsukamoto, H., 188 Turke-Browne, N. B., 257 Turner, R. S., 11, 12 Turuani, L., 45 Turvey, M. T., 355 Ungerleider, L. G., 107–109 Valbo, A., 406 Van der Gucht, E., 411 Van der Helm, F. C. T., 16 Van der Kamp, J., 219 Van der Schans, C. P., 427 Vandewalle, G., 421 Van Essen, D. C., 107 Van Gemert, L. J., 444 Van Opstal, A. J., 325 Van Paassen, M. M., 16 Van Riswick, J. G., 325 Van Zandt, T., 180 Vaso, A., 428 Vasterling, J. J., 18 Vaughan, J. T., 281 Venrooij, J., 16 Verstraten, F. A. J., 10 Victorsson, J., 56 Vidoni, E. D., 146–147 Vo, M. L.-H., 264, 266 Von Arx, S. W., 175 Von Cramon, J. D., 239 Von der Heydt, R., 135 Voss, P., 421 Vuilleumier, P., 277 Vuvan, D. T., 387 Vyshka, G., 428 Wade, A., 175 Wade, N. J., 10, 69 Wagener, K. C., 48 Wake, H., 419 Wald, G., 67 Wald, M. L., 125 Walenchok, S. C., 146–147 Walker, L., 454 Walker, T., 241 Wall, P. D., 410, 415 Walls, G. L., 65 Walsh, E., 427 Walsh, V., 270 Walter, S., 20 Wandell, B., 175 Wang, C., 233 Wang, Y.-H., 117 Wang, Y.-U., 283 Wang duncan dietrich, 456 Wang schoenlein peteanu mathies andshank, 67 Wann, J. P., 22 Ward, J., 386 Ward, L., 418 Warren, R. M., 354 Warren, W. H., 244 Warrington, E. K., 114 Warzybok, A., 365 Washburn, S., 420 Wasserman, L., 68, 174 Author Index 499 Watanabe, I., 211 Watanabe, M., 73 Watson, J. M., 256 Watson, R., 361 Weeks, S. R., 427 Weigelt, S., 238 Weinberg, A., 144 Weiskopf, S., 168 Weiskrantz, L., 18, 114–115, 281 Weitzmann, K., 10 Welburn, S. C., 282 Welsh-Bohmer, K. A., 454 Werker, J. F., 357 Wert, K. J., 81 Wertheimer, M., 132–133 Wessberg, J., 406 Westhoff, C., 241 Wever, E. G., 305 Wheatley, T., 143–145, 144, 241 White, R. F., 18 White-Schwoch, T., 363 Wickens, C. D., 146–147 Wiesel, T. N., 17, 91, 103–104, 112–113, 209 Wilbiks, J. M. P., 387 Wilcox, K. J., 247 Williams, D. R., 171 Williams, S. J., 409 Willmes, K., 385 Wilson, D. A., 442 Wilson, K., 215 Wirantana, V., 384 Witherby, S. A., 386 Witt, R., 428 Wolfe, J. M., 264, 267 Wolford, G. L., 19 Wong-Riley, M. T., 105 Wood, M. L., 416 Woods, C. B., 170 Wright, E., 418 Wurtz, R. H., 235 Xiao, M., 168 Xiao, Y., 97 Xie, W.-J., 117 Xu, X., 97 Xue, J., 42 Xu-Friedman, M. A., 318 Yamamoto, H., 354 Yamane, H., 354 Yang, H., 318 Yang, J. S., 42 Yang, T. T., 456 Yantis, S., 265, 273, 274 (figure) Yao, H., 233 Yarbrough, G. L., 452 Yasuaki, T., 250 Yee, C. W., 452 Yeh, Y.-Y., 72 Yin, T. C., 323 Yoneyama, K., 215 Yoshida, R., 450 Yost, W. A., 307 Yovel, G., 110 Yu, H.-Q., 117 Yu, R., 20 Zahaj, S., 428 Zakis, J. A., 395 Zamm, A., 386 Zanos, T. P., 259 Zatorre, R. J., 384, 421, 445 Zeil, J., 83 Zeki, S., 106, 109, 111, 239 Zelterman, D., 418 Zhang, B., 211 Zheng, J., 211 Zhou, W., 446 Zhu, M., 282 Zhurda, T., 428 Zihl, J., 239 Zimba, L. D., 260, 261 Zohary, E., 237 Zokoll, M. A., 48 Zosh, W. D., 244 Zucco, G. M., 45 Zuker, C. S., 450 Sensation and Perception500 SUBJECT INDEX Absolute threshold, 30, 31 (figure) stars as, 31, 31 (figure) Accommodation (eye), 60–61, 199 Accretion, 197, 198 (figure) Acoustic reflex, 301 Acoustic shadow, 324, 324 (figure) Acoustics of music consonance and dissonance in, 376–377 dynamics and rhythm in, 377–378, 378 (figure) melody in, 380–383, 380 (figure), 382–383 (figure) pitch, chroma, and the octave in, 373–376, 374–376 (figure) timbre in, 378–379, 379 (figure) Acoustics of phonemes, 347–348 (figure), 347–355, 350–353 (figure) Action, 8, 8 (figure), 242–246 Active electroreception, 425–426 Acuity, visual, 70 infant, 72–73, 73 (figure) Adaptation, dark and light, 70–72, 71 (figure), 72 (figure) Additive color mixing, 157–158, 157–158 (figure) A-delta fibers, 409 Afferent fibers, 407 Affordances, 242–246, 243 (figure) Aftereffects, 10, 10 (figure) motion, 239–240, 240 (figure) Afterimages, 164–165 Ageusia, 455 Aging basilar membrane of ear and, 306 color perception and, 169 hearing loss with, 306–307 Agnosia, 18 object, 109, 124, 139 tactile, 421–422 topographic, 140 Airplane pilots, motion perception in, 249–250 Airport screening, 145–147, 146 (figure), 270 (figure) Akinetopsia, 239 Ames room illusion, 215–216, 216 (figure) Amplitude and loudness, 292–294, 293–294 (figure) Ampulla, 423 Ampullae of Lorenzini, 426 Amusia, 18 Amygdala, 441 Analgesia, 417 Animacy, 143–145, 143–145 (figure) Animals affordances and, 242–246, 243 (figure) awareness in, 279 (figure) biosonar in, 331–334, 332–334 (figure) electroreception in, 425–426, 426 (figure) eyes of, 82–83 (figure), 82–84 Anomalous trichromacy, 173–174, 174 (table) Anosmia, 440, 453–455 Anterior chamber of eye, 60 Anterior cingulate cortex, 416 Anterior insular cortex, 450 Anterior intraparietal (AIP) area, 246 Anterior piriform cortex, 442 Aphasia, 359 Apparent motion, 230–231 (figure), 230–232 Architectural acoustics, 335 Aristotle, 10 Aristotle illusion, 10, 10 (figure) Armstrong, Louis, 378, 378 (figure) Artificial sweeteners, 455–457, 456 (figure) Ascending series, 31 Astigmatism, 79–80, 80 (figure) Atmospheric perspective, 194–195, 195 (figure) Attack, 379 Attention anatomy and physiology of, 272–277, 273–277 (figure) attentional blink and rapid serial visual presentation, 270–272, 271 (figure) awareness and visual consciousness, 279–282, 279–282 (figure) covert, 259–260, 260 (figure) defined, 257 developmental aspects of, 278–279 and direction of gaze in space, 259–262, 259–262 (figure) divided, 258, 258 (figure) effects on the visual brain, 274–275, 275 (figure) feature integration theory, 267, 268 (figure) inattentional blindness and, 262–264, 263–264 (figure), 265 (figure) infants and, 278–279, 278–279 (figure) introduction to, 255–257, 256–257 (figure) neuropsychology of, 275–277, 276–277 (figure) over time, 268–272, 269–271 (figure) selective, 258 spatial limits of, 259–264, 259–264 (figure) spotlight model of, 261, 262 (figure) stimulus features that draw, 264–267, 265–266 (figure) visual search and, 266–267, 267 (figure), 268 (figure) Attentional blink, 270–272, 271 (figure) Attentional capture, 265, 265 (figure), 266 (figure) Audiograms, 47, 47 (figure) Audiologists, 46, 47 (figure) Audiometers, 47 Auditory brain anatomy and pathway of hearing in, 318–321, 319–321 (figure) auditory scene analysis, 326–330, 327–331 (figure) introduction to, 317–318, 318 (figure) localizing sound, 321–326, 322–326 (figure) neuroanatomy of music and, 384–385, 384–385 (figure) speech perception and, 358–362, 359–362 (figure) Auditory core region, 320 Auditory cortex, 320–321, 321 (figure) Auditory development, 330 Auditory nerve fibers, 318–320, 319–320 (figure) Auditory system, 7, 289–290, 290 (figure) anatomy of the ear and, 298–306, 298–306 (figure) hearing loss and, 306–308 sound as stimulus and, 290–297, 290–297 (figure) tests of, 45–49 See also Auditory brain Auto collisions, 21–22 Automaticity, 258 Awareness and visual consciousness, 279–282, 279–282 (figure) Axel, Richard, 439–440 Azimuth, sound, 322 Balance, perception of, 423–425, 424 (figure) Bálint’s syndrome, 277 Basal cells, 438 Basilar membrane of the cochlea, 302–304, 302–304 (figure) Bats, biosonar in, 331–334, 332–334 (figure) Beat, 378 Bell, Joshua, 255–256, 256 (figure) Belt and parabelt regions, 320 Benham’s top, 13, 13 (figure) Biases, personal, 3–4 Binocular cells, 209–210 Binocular cues to depth, 200–201 Binocular disparity, 201, 201 (figure) Binocular rivalry, 280–281, 280 (figure) Biosonar in bats and dolphins, 331–334, 332–334 (figure) Bistability, 280–281, 280–282 (figure) Bistratified retinal ganglion cells (K cells), 95 Blindness change, 268–270, 269–270 (figure) inattentional, 262–264, 263–264 (figure), 265 (figure) repetition, 272 smell, 440 Blindsight, 113–116, 115 (figure), 282, 282 (figure) Blobs, V1, 105 Bottom-up processing, 126–128, 127–128 (figure) Braille, 420–421, 420 (figure) Braille, Louis, 420 Brain, auditory anatomy and pathway of hearing in, 318–321, 319–321 (figure) auditory scene analysis, 326–330, 327–331 (figure) introduction to, 317–318, 318 (figure) Subject Index 501 localizing sound, 321–326, 322–326 (figure) neuroanatomy of music and, 384–385, 384–385 (figure) speech perception and, 358–362, 359–362 (figure) Brain, visual, 91–92 anatomy and physiology of attention and, 272–277, 273–277 (figure) blindsight and, 113–116, 115 (figure) conjugate gaze palsy and, 116–117 damage and neuropsychology, 18 development of vision and, 111–113 executive attention network, 273–274, 274 (figure) functional pathways in visual cortex of, 107–110, 108–109 (figure) how attention affects, 274–275, 275 (figure) lateral geniculate nucleus, 94–98, 95–97 (figure) lobes, 100, 100 (figure) mapping the eye on the, 101–102, 101–102 (figure) MT as movement area of, 236–239, 237–239 (figure) object perception in, 137–142, 137–142 (figure) optic nerve and chiasm, 92–94, 93 (figure) orienting attention network, 273, 273 (figure) primary visual cortex, 99–106 superior colliculus, 98–99, 98 (figure) V2 and beyond, 106–107, 106 (figure) where vision comes together in, 111 Brain pathways in olfaction, 440–443 in pain, 410–416, 411–416 (figure) Brightness, 156 Broca, Pierre, 359, 359 (figure) Broca’s aphasia, 359 Broca’s area, 359 Buck, Linda, 439–440 Capsaicin, 27, 28 (figure) Cataracts, 80 Catch trials, 35–36 Categorical perception, 349–351, 350 (figure) Cats affordances and, 242–246, 243 (figure) awareness in, 279 (figure) eyes of, 82–83, 82–83 (figure) Center-surround receptive fields, 74, 75 (figure) C-fibers, 409 Change blindness, 268–270, 269–270 (figure) Characteristic frequency, 303 Chili peppers, 452–453, 452 (figure) Chroma, 375 Ciliary muscles, 61 Circumvallate papillae, 449 Closure (music), 383, 383 (figure) Coarticulation, 348–349, 348 (figure) Cochlea, 302, 302 (figure) basilar membrane of, 302–304, 302–304 (figure) Cochlear duct, 302 Cochlear implants, 308–311, 309–311 (figure) Cochlear nucleus, 319 Cognitive impenetrability, 20 Cognitive penetration, 19–21 Cold fibers, 408 Collisions, 21–22 Color, complementary, 165 Color circle, 156, 156 (figure) Color constancy, 175–177, 175–177 (figure) Color deficiency, 169–175, 170 (figure), 172 (table), 173 (figure), 174 (table) Color-matching experiments, 158–159, 159 (figure) Color-music synesthesia, 386, 386 (figure) Color-opponent cells, 167, 167 (figure) Color perception, 151–153 additive and subtractive color mixing and, 156–160, 157–159 (figure) aging and, 169 constancy in, 175–178, 175–178 (figure) deficiency, 169–175, 170 (figure), 172 (table), 173 (figure), 174 (table), 180–181, 180–181 (figure) development of, 168–169, 168 (figure) hue, saturation, lightness, and brightness in, 155–156, 155–156 (figure) opponent-process theory of, 163–167, 163–167 (figure) purple, 178–179, 179 (figure) retina and, 160–162, 161 (figure) trichromatic theory of color vision and, 162–163, 163 (figure) wavelengths of light and color and, 153–155, 154 (figure) Comic books, 129, 129 (figure) Complementary colors, 165 Complex cells, V1, 103–104, 103–104 (figure) Complex sounds, 296 Compound eye, 83–84, 84 (figure) Computational approach, 16–17 Computerized tomographic (CT) scans, 42 Computer speech recognition, 364–366, 365 (figure) Concert hall acoustics and hearing, 335–336, 335 (figure) Conductive hearing loss, 46, 307 Cone of confusion, 324, 325 (figure) Cone-opponent cells, 166–167 Conjugate gaze palsy, 116–117 Conjunction search, 267 Consonance and dissonance, 376–377 Consonants and vowels, 343–344, 344 (figure) Constructivist approach, 11 Continuation (music), 383 Contralateral organization, 93 Contralateral representation of visual space, 93 Contralateral side, 412 Convergence, 70, 199–200, 200 (figure) Convexity, 132, 132 (figure) Cornea, 59, 59 (figure) Corollary discharge theory, 234–235, 234–235 (figure) Correct rejection, 37 Correspondence problem, 206 motion and, 231, 231 (figure) Corresponding points, 202–206, 203–205 (figure) Cortical achromatopsia, 174–175 Cortical magnification, 101 Covert attention, 259–260, 260 (figure) Cribriform plate, 440 Crista, 423 Criterion, 37, 40 Crossed disparity, 203 Crossover point, 31 Cue approach to depth perception, 189 Culture and music perception, 389–391, 390–391 (figure) Cycles, 291 Dark adaptation, 70–72, 71 (figure), 72 (figure) Decay, 379 Decibels (dB), 293–294 Deletion, 197, 198 (figure) Depth, visual illusions of, 213–217 Depth perception binocular cues to, 200–201 cue approach to, 189 introduction to, 187–189, 188 (figure) monocular depth cues to, 189–199, 190–198 (figure) oculomotor cues to, 199–200, 200 (figure) stereograms and, 207–210 (figure), 207–211 Dermis, 403 Descending series, 31 Detection of distance, 325–326 olfactory, 443–444 Deuteranopia, 172–173 Dichromacy, 172–174, 173 (figure), 174 (table), 180–181, 180–181 (figure) Difference threshold, 30 Diplopia, 203 Direct perception, 15 Dishabituation, 168 Dissonance and consonance, 376–377 Distance, sound, 322 detecting, 325–326 Distracted driving, 256 (figure), 282–283 Divided attention, 258, 258 (figure) Doctrine of specific nerve energies, 11 Dolphins, biosonar in, 331–334, 332 (figure) Doppler shifts, 333, 334 (figure) Dorsal attention network, 273, 273 (figure) Dorsal column-medial lemniscal pathway, 412 Dorsal horn, 415 Dorsal pathway, 107, 109 (figure), 110 Dorsal root ganglion, 411–412, 411 (figure) Double-opponent cells, 167 Driving, distracted, 256 (figure), 282–283 Dual-stream model of speech perception, 360–361, 360 (figure) Dual-target cost, 146 Duplex theory of vision, 68–73 Dynamics and rhythm, 377–378, 378 (figure) Sensation and Perception502 Ear(s) anatomy of, 298–306, 298–306 (figure) basilar membrane of the cochlea, 302–304, 302–304 (figure) inner, 301–302, 301 (figure) middle, 299–301, 300 (figure) organ of Corti, 303, 304–306, 305 (figure) outer, 298, 299 (figure) piercing of, 325, 326 (figure) protection of, 290 (figure) See also Auditory system Ebbinghaus, Hermann, 13 Ecological approach to perception, 15 Edge completion, 134–135, 134 (figure) Edge detection/enhancement, 74, 76–77, 77 (figure) Electroencephalography (EEG), 43 Electromagnetic energy, 55 Electromagnetic spectrum, 56, 56 (figure) Electroreception in fish, 425–426, 426 (figure) Elevation, sound, 322 perception, 325, 325 (figure) Endogenous opioids, 417 Endolymph, 423 End-stopped neurons, 104 Entorhinal cortex, 441 Epidermis, 403 Equal-temperament scale, 376 Eustachian tube, 300 Executive attention network, 273–274, 274 (figure) External auditory canal, 298 External auditory meatus, 298 Extrastriate body area, 140 Eye(s), 57–62 anatomy of, 58–61, 59–61 (figure) in animals, 82–83 (figure), 82–84 emerging and aging of, 61–62 field of view of, 58, 58 (figure) lens of, 60–61, 61 (figure) mapping on the brain, 101–102, 101–102 (figure) movements of, 235–236 refractive errors and diseases of, 77–81 retina of, 59 (figure), 62–68, 62 (figure) vision prostheses for, 85–86, 86 (figure) visually guided movements of, 244–245, 245 (figure) See also Visual system Face agnosia, 18 FAII mechanoreceptors, 404, 405 (figure), 406 FAI mechanoreceptors, 404, 405–406, 405 (figure) False alarms, 37, 37 (figure), 41 (figure), 42 Familiar size, 192–193, 192–193 (figure) Farsightedness, 78–79, 79 (figure) Feature integration theory, 267, 268 (figure) Feature search, 267 Fechner, Gustav, 12–13, 12 (figure) Field of view, 58, 58 (figure) Figure-ground organization, 129–131, 129–131 (figure) Filiform papiallae, 449 First cranial nerve, 441 Fish, electroreception in, 425–426, 426 (figure) Fisher, M. F. K., 447 Five-sense approach, 4–5, 4 (table) Flavor, 447, 451, 451 (figure) See also Taste Foliate papillae, 449 Forced-choice method, 36, 36 (figure) Formants, 343–344, 344 (figure) Form perception and biological motion, 241–242, 241–242 (figure) Fourier analysis, 296 Fovea, 64–65 Frequency, wave, 55 Frequency and pitch, 294–295, 295 (figure) Functional MRI (fMRI), 18–19, 19 (figure), 43 Fundamental frequency, 296 Fungiform papillae, 449 Furrow illusion, 248–249, 249 (figure) Fusiform face area (FFA), 110, 138–139, 138 (figure), 139 (figure) Gate control theory, 415, 416 (figure) Gaze in space, 259–262, 259–262 (figure) Gelb effect, 178 General-mechanism theories, 355–356 Genes olfaction and, 439–440 taste perception and, 451 Geons, 136, 136 (figure) Gestalt psychology, 13–15 figure-ground organization in, 129–131, 129–131 (figure) laws of, 14 (figure) laws of perceptual grouping, 132–134, 132–134 (figure) melody and, 381–383 perceptual interpolation and, 134–135, 134–135 (figure) perceptual organization and, 128–135, 129–135 (figure) rules that govern what we see as figure and what we see as ground in, 131–132, 131–132 (figure) Gibsonian approach, 15 Glomeruli, 441 Golgi tendon organs, 407 Grandmother cells, 140–142 Grasping, visually guided, 245–246 Grouping, 128, 128 (figure) laws of perceptual, 132–134, 132–134 (figure) Gustation, 436 Habituation, 168 Hair cells, 304 Hand clapping, 290–291, 291 (figure) Haptic reception, 419–420 (figure), 419–422 Harmonic coherence, 329 Harmonics, 296 Harmony, 377 Hearing aids, 308–311, 309–311 (figure) Hearing loss, 306–308 conductive, 46, 307 music perception with, 394–396, 395–396 (figure) sensorineural, 46, 307–308 speech perception and, 362–364 Helmholtz, Hermann von, 11–12, 11 (figure), 162–163, 163 (figure) Hemifield neglect, 276–277, 277 (figure) Hering, Ewald, 12, 12 (figure), 164 Hertz (Hz), 294–295 Heterochromatic light, 153 Heterochromia, 60 Hit, 37 Homunculus, 414 Hooke, Robert, 11 Horopter, 202–203, 204 (figure) Hubel, David, 91–92, 92 (figure) Hue, 155–156 cancellation, 165–166, 166 (figure) Human voice as stimulus, 342–347, 343–345 (figure), 346 (table) Hyperocolumns, V1, 105, 105 (figure) Hyperopia, 78–79, 79 (figure) Hyperpolarization, 67 Illusions motion, 246–249, 247–249 (figure) musical, 391–394, 392–394 (figure) octave, 392, 393 (figure) olfactory, 445–446 scale, 393, 393 (figure) Illusory contours, 135, 135 (figure) Illusory rotation, 248, 248 (figure) Imagery, odor, 445 Inattentional blindness, 262–264, 263–264 (figure), 265 (figure) Incus, 299 Induced motion, 231–232 Infants attention in, 278–279, 278–279 (figure) auditory development in, 330 color perception in, 168–169, 168 (figure) phoneme perception in, 357–358 stereopsis in, 210–211 taste development in, 453, 453 (figure) visual acuity in, 72–73, 73 (figure) Inferior colliculus, 319 Inferotemporal area, 137–138, 137 (figure) Inferotemporal cortex, 109, 140–142 Information-processing approach, 15–16, 16 (figure) Inner ear, 301–302, 301 (figure) Inner hair cells, 304 Insula, 450 Intensity, wave, 55 Interaural level difference, 324, 324 (figure) Interaural time difference, 322–323, 323 (figure) Interblobs, V1, 105 International Phonetic Alphabet, 346–347, 346 (table) Internuclear ophthalmoplegia, 117 Interpolation, perceptual, 134–135, 134–135 (figure) Intersensory perception, 44–45, 45 (figure) Ipsilateral organization, 93 Subject Index 503 Iris of eye, 60, 60 (figure) Ishihara plate, 170, 170 (figure) Itch, 418, 418 (figure) James, William, 13 Joint receptors, 406–407 Just-noticeable difference (JND), 12–13, 28, 30 Kanizsa triangle, 13–14, 14 (figure) Keys and scales, 380–381 Knowledge influence on perception and sensation, 4 Koffka, Kurt, 128 Köhler, Wolfgang, 15, 15 (figure), 128 Koniocellular layers, 94–95 Koniocellular pathway (K pathway), 96 Kuffler, Stephen, 74, 82, 91–92 Language and music, 388–389 Larynx, 343, 343 (figure) Lateral geniculate nucleus, 94–98, 95–97 (figure) opponent cells in, 166–167, 167 (figure) Lateral inhibition, 76–77 Lateral intraparietal (LIP) area, 245, 245 (figure) Law of closure, 14 (figure) Law of common fate, 14 (figure), 134 Law of good continuation, 14 (figure), 133, 133 (figure) Law of proximity, 14 (figure), 133 Law of similarity, 14 (figure), 133, 133 (figure) Law of symmetry, 133–134, 134 (figure) Laws of perceptual grouping, 132–134, 132–134 (figure) L-cone, 160–161, 161 (figure) Lens of eye, 60–61, 61 (figure) Light, 55–57, 57 (figure) adaptation, 71–72 color and wavelengths of, 153–155, 154 (figure) mixing, 157–158, 157–158 (figure) transduction of, 6, 66–67 Lightness, 156 Lightness constancy, 177–178, 178 (figure) Linear perspective, 193–194, 193–194 (figure) Localizing of sound, 321–326, 322–326 (figure) Loudness and amplitude, 292–294, 293–294 (figure) Mach, Ernst, 76 Mach bands, 76, 77 (figure) Macrosmatic species, 439 Macula, 64, 423 Macular degeneration, 80, 81 (figure) Magnetic resonance imaging (MRI), 18, 43, 43 (figure) Magnetoencephalography (MEG), 42 (figure), 43 Magnitude estimation, 34–35, 35 (table) Magnocellular layers, 94–95 Magnocellular pathway (M pathway), 96 Malleus, 299 Manner of articulation, 344 “Mary Had a Little Lamb,” 381 McGurk effect, 351–352, 351–352 (figure) M-cone, 160, 161 (figure) Mechanoreception, 404–406, 405 (figure) Medial geniculate nucleus, 319 Medial intraparietal (MIP) area, 246 Melody, 380–383, 380 (figure), 382–383 (figure) Metamer, 159, 159 (figure) Meter, 378 Method of adjustment, 33–34 Method of constant stimuli, 32–33, 33 (figure) Method of limits, 29–32, 30 (figure) Microelectrodes, 17 Microsomatic species, 439 Middle canal, 302 Middle ear, 299–301, 300 (figure) Midget retinal ganglion cells (P cells), 95 Mitral cells, 441 Monochromacy, rod and cone, 171–172, 172 (table) Monochromatic light, 153 Monocular depth cues, 189–199, 190–198 (figure) atmospheric perspective, 194–195, 195 (figure) familiar size, 192–193, 192–193 (figure) linear perspective, 193–194, 193–194 (figure) motion cues, 196–199, 197–198 (figure) occlusion, 190, 190 (figure) relative height, 191, 191 (figure) relative size, 191–192 shadows and shading, 195–196, 196 (figure) texture gradients, 194, 194 (figure) Moon illusion, 216–217, 217 (figure) Motion aftereffects, 239–240, 240 (figure) airline pilots and perception of, 249–250 complexity of, 232–233, 233–234 (figure) corollary discharge theory, 234–235, 234–235 (figure) defined, 228 eye movements, 235–236 form perception and biological, 241–242, 241–242 (figure) how we perceive, 228–232, 229–231 (figure) illusions of, 246–249, 247–249 (figure) induced, 231–232 MT as movement area of the brain and, 236–239, 237–239 (figure) neuroscience of vision and, 232–240, 233–235 (figure), 237–240 (figure) real and apparent, 230–231 (figure), 230–232 Motion blindness, 239 Motion cues, 196–199, 197–198 (figure) Motion parallax, 196–197, 197 (figure) Motion thresholds, 228–229, 229 (figure) Motor theory of speech perception, 355 Movement-based cues, 189, 190 (figure) MT (middle temporal) region, 110 as movement area of the brain, 236–239, 237–239 (figure) Mueller, Johannes, 11 Müller-Lyer illusion, 214–215, 214 (figure), 215 (figure) Muscle spindles, 406 Music acoustics of, 373–383, 374–376 (figure), 378–380 (figure), 382–383 (figure) introduction to, 371–373, 371–373 (figure) language and, 388–389 learning, culture, and perception of, 388–391 neuroanatomy of, 384–385, 384–385 (figure) neuropsychology of, 386–388, 387 (figure) neuroscience of, 383–388, 384–387 (figure) perception in hearing-impaired listeners, 394–396, 395–396 (figure) synesthesia, 385–386, 386 (figure) Musical illusions, 391–394, 392–394 (figure) Musical notation, 380, 380 (figure) Myopia, 48–49, 78, 78 (figure) Narrowing, perceptual, 357–358 Nasal septum, 437–438 Nautilus, 83, 83 (figure), 84 (figure) Near point (eye focus), 61, 61 (figure) Nearsightedness, 48–49, 78, 78 (figure) Necker cube, 135, 135 (figure) Neural pathways of pain, 410–416, 411–416 (figure) Neural response, 6–7, 7 (figure) Neurochemistry of pain, 417 Neuroimaging, 18 Neuroimaging techniques, 42–43, 42–43 (figure) Neuropsychology, 18 of attention, 275–277, 276–277 (figure) of music, 386–388, 387 (figure) Neuroscience of music, 383–388, 384–387 (figure) in sensation and perception, 17–19 of vision and motion, 232–240, 233–235 (figure), 237–240 (figure) Neurotransmitters, 67 Nociception, 409–410, 410 (figure) Nociceptors, 409 Noncorresponding points, 202–206, 203–205 (figure) Nontasters, 451, 452 (figure) Nose, 437–439, 438–439 (figure) Object agnosia, 109, 124, 139 Object perception, 123–124 airport screening and, 145–147, 146 (figure) animacy and, 143–145, 143–145 (figure) gestalt psychology, 128–135, 129–135 (figure) introduction to, 124–126, 124–126 (figure) neuroanatomy and physiology of, 137–142, 137–142 (figure) Sensation and Perception504 recognition by components, 136, 136 (figure) top-down processing and bottom-up processing, 126–128, 127–128 (figure) Occipital face area (OFA), 138, 138 (figure) Occlusion, 190, 190 (figure) Octave, 375, 375 (figure) Octave illusion, 392, 393 (figure) Ocular dominance columns, V1, 104 Oculomotor cues, 199–200, 200 (figure) Oddball procedure, 278 Odors, 436, 437 (figure) identifying, 444–445, 444–445 (figure) imagery, 445 Off-center receptive fields, 75 Old-sightedness, 61, 61 (figure), 62, 78–79 Olfaction, 436, 437 (figure) anatomy and physiology, 437–443 brain pathways, 440–443 genes and, 439–440 identifying odors, 444–445, 444–445 (figure) illusions, 445–446 introduction to, 435–436 perception, 443–446 trigeminal nerve, 440 Olfactory bulb, 441, 442 (figure) Olfactory cleft, 438 Olfactory epithelium, 438, 439 (figure) Olfactory nerve, 441 Olfactory receptor neurons, 438 Olfactory tract, 441 On-center receptive fields, 75 Opioids, endogenous, 417 Opponent-process theory of color perception, 163–167, 163–167 (figure) Opsins, 66, 66 (figure) Optic disc, 65 Optic flow, 198–199, 198 (figure) Optic nerve and chiasm, 92–94, 93 (figure) Optic tract, 93 Optometrists, 29, 29 (figure), 48 Orbitofrontal cortex, 443 Organ of Corti, 303, 304–306, 305 (figure) Orientation columns, V1, 104 Orienting attention network, 273, 273 (figure) Orienting tuning curves, 103 Ossicles, 299–300 Otolith organs, 423 Otosclerosis, 307 Outer ear, 298, 299 (figure) Outer hair cells, 304 Pain. See Touch and pain Paints, mixing, 158, 158 (figure) Paley, William, 58 Panum’s area of fusion, 203 Papillae, 449 Parahippocampal place area (PPA), 140, 140 (figure) Parasol retinal ganglion cells (M cells), 95 Parietal insular vestibular cortex, 425 Parvocellular layers, 94 Parvocellular pathway (P pathway), 96 Passive electroreception, 425 Patient T. N., 113–116 Penfield, Wilder, 413 Perception basics of, 5–9 biases in, 3–4 categorical, 349–351, 350 (figure) color (See Color perception) defined, 5 depth (See Depth perception) differentiated from sensation, 5–6 direct, 15 elevation, 325, 325 (figure) history of, 9–23 intersensory, 44–45, 45 (figure) introduction to, 1–3 knowledge influence on, 4 music (See Music) myth of the five senses and, 4–5, 4 (table) neuroscience in, 17–19 object (See Object perception) as part of psychology, 3–4 size, 211–212 speech (See Speech perception) Perceptual bistability, 280–281, 280–282 (figure), 280 (figure) Perceptual interpolation, 134–135, 134–135 (figure) Perceptual narrowing, 357–358 Perceptual organization, 128, 128 (figure) gestalt psychology and, 128–135, 129–135 (figure) Perilymph, 303 Phantom limbs, 427–428 Phantosmia, 455 Pharynx, 343, 343 (figure) Phase, sound, 297, 297 (figure) Phenomenology, 8–9, 8 (figure) Phonemes, 345, 346 (table) perception, development of, 357–358 variability in the acoustics of, 347–348 (figure), 347–355, 350–353 (figure) Phonemic restoration effect, 353–355 Photons, 56–57 Photopic system, 68, 69 (table) Photopigments, 66–67, 67 (figure) Pictorial cues, 189, 190 (figure) Piercing, ear, 325, 326 (figure) Pinna, 298, 299 (figure) Piriform cortex, 441, 442 Pitch, 373–376, 374–376 (figure) frequency and, 294–295, 295 (figure) Pitch class circle, 394, 394 (figure) Place code theory, 305 Place of articulation, 344 Point-light walker display, 241, 242 (figure) Point of subjective equality (PSE), 33 Ponzo illusion, 215, 215 (figure) Pop-out search, 267 Posner cuing paradigm, 261, 261 (figure) Posterior chamber of eye, 60 Posterior piriform cortex, 442 Presbycusis, 306–307 Presbyopia, 61, 61 (figure), 62, 78–79 Presynaptic cells, 449 Primary auditory cortex, 320–321, 321 (figure) Primary visual cortex, 99–106 Private experience, phenomenology as, 9 Proprioception, 406–407, 407 (figure) Prosopagnosia, 18, 139–140 Prostheses, vision, 85–86, 86 (figure) Protanopia, 172, 173 (figure) Proximity (music), 382 Pruriceptors, 418 Psychology gestalt, 13–15, 128–135, 129–135 (figure) information-processing approach in, 15–16 sensation and perception as, 3–4 Psychophysical scales, 27–29 catch trials and their use, 35–36 magnitude estimation, 34–35, 35 (table) method of adjustment, 33–34 method of constant stimuli, 32–33, 33 (figure) method of limits, 29–32 neuroimaging techniques, 42–43, 42–43 (figure) signal detection theory, 36–37, 36–42, 37–42 (figure), 37 (figure), 37 (table), 39 (table) Psychophysics, 12–13 in assessment, 45–49 discrimination in, 444–445 measures and methods of, 29–49 scales, 27–29 Pupillary reflex, 60, 60 (figure) Pupils, 60 Pure tones, 292, 292 (figure) Purkinje shift, 69–70 Purple (color), 178–179, 179 (figure) Quality (of hue), 155 Random-dot stereograms, 208–209, 209 (figure) Rapid serial visual presentation (RSVP), 270–272, 271 (figure) Rate of approach, 333 Real and apparent motion, 230–231 (figure), 230–232 Receiver-operating characteristic (ROC) curve, 41–42, 41 (figure) Receptive fields (vision), 73–75 (figure), 73–77, 77 (figure) Receptors, 6, 68, 69 (table), 449 Recognition, 127, 127 (figure) by components, 136, 136 (figure) Reichardt detectors, 233, 234 (figure) Reintal (vitamin A), 66–67 Reissner’s membrane, 302 Relative height, 191, 191 (figure) Relative size, 191–192 Repetition blindness, 272 Representation, 127, 127 (figure) Response compression, 34 Response expansion, 34–35 Retina, 59 (figure), 62–68, 62 (figure) anatomy of, 63, 63 (figure), 64 (table) color and, 160–162, 161 (figure) Subject Index 505 motion detection in, 232 physiology, 65–68 receptors, 64, 68, 69 (table) Retinal ganglion cells, 73–75 (figure), 73–77, 77 (figure) Retinal image, 62 Retinitis pigmentosa, 80–81, 81 (figure) Retinoptic map, 101, 101 (figure) Reverberation time, 335 Rhythm and dynamics, 377–378, 378 (figure) Robots, 9 Rods and cones of eye, 64–65, 65 (figure) monochromacy, 171–172, 172 (table) wavelength sensitivity of, 68 (figure) Rostral core, 320 Rostrotemporal core, 320 Rotating snakes illusion, 247–248, 247 (figure) Round window, 302 Saccades, 99, 236 SAII mechanoreceptors, 404, 405, 405 (figure) SAI mechanoreceptors, 404, 405, 405 (figure) Salience, stimulus, 265–266 Saturation, 155–156, 155 (figure) Scale illusion, 393, 393 (figure) Scales and keys, 380–381 Scanning, 259, 259 (figure) Sclera, 60 S-cone, 160, 161 (figure) Scotoma, 114–115 Scotopic system, 68, 69 (table) Scoville, Wilbur, 28 Scoville scale, 27–28, 28 (figure) Search, visual, 266–267, 267 (figure), 268 (figure) Segregation, 128, 128 (figure) Selective attention, 258 Semicircular canals, 423 Semitones, 375 Sensation biases in, 3–4 defined, 5 differentiated from perception, 5–6 history of, 9–23 introduction to, 1–3 knowledge influence on, 4 myth of the five senses and, 4–5 neuroscience in, 17–19 as part of psychology, 3–4 Senses, myth of five, 4–5, 4 (table) Sensitivity, 34, 40 (figure), 41–42 Sensitivity of detection, 44–45, 45 (figure) Sensorineural hearing loss, 46, 307–308 Sensory neurons, 17, 17 (figure) Shadows and shading, 195–196, 196 (figure) Shepard tones, 392, 392 (figure) Signal detection theory, 36–42, 37–42 (figure), 37 (table), 39 (table) Sign language, 306, 307 (figure) Similarity (music), 382 Simple cells, V1, 103, 103 (figure) Simultagnosia, 277 Simultaneous color contrast, 165, 165 (figure) Size, visual illusions of, 213–217 Size-arrival effect, 22 Size constancy, 212–213, 213 (figure) Size-distance invariance, 211–212 Size perception, 211–212 Skin, 403, 403 (figure) mechanoreception, 404–406, 405 (figure) proprioception, 406–407, 407 (figure) thermoreception, 407–409, 408 (figure) Smell blindness, 440 Smooth-pursuit eye movements, 98–99, 235 (figure), 236 Snellen chart, 48, 48 (figure) Sobriety testing, 407, 407 (figure) Somatosensory cortex, 413–414, 413–414 (figure) suborganization of, 414–415, 415 (figure) Somatotopic map, 414, 414 (figure) Sound amplitude and loudness of, 292–294, 293–294 (figure) biosonar in bats and dolphins, 331–334, 332–334 (figure) complex, 296 concert hall acoustics and hearing, 335–336, 335 (figure) cone of confusion, 324, 325 (figure) elevation perception, 325, 325 (figure) frequency and pitch, 294–295, 295 (figure) interaural level difference, 324, 324 (figure) interaural time difference, 322–323, 323 (figure) localizing, 321–326, 322–326 (figure) phase, 297, 297 (figure) phonemic restoration effect, 353–355 relation of physical and perceptual attributes of, 292, 292 (figure) spatial segregation, 329 spectral segregation, 329–330, 330 (figure), 331 (figure) speed of, 291–292, 292 (figure) as stimulus, 290–297, 290–297 (figure) temporal segregation, 328–329, 328 (figure) top-down processing, 126–128, 127–128 (figure), 352–353, 353 (figure) waveform and timbre, 295–297, 296 (figure) See also Auditory system Sound waves, 55, 291, 291 (figure) Spatial limits of attention, 259–264, 259–264 (figure) Spatial segregation, 329 Spatial summation and acuity, 70 Special-mechanism theories, 355 Spectral reflectance, 153 Spectral segregation, 329–330, 330 (figure), 331 (figure) Spectral sensitivity, 69–70 Spectral shape cue, 325 Speech, 344–347, 345 (figure), 346 (table) computer recognition of, 364–366, 365 (figure) Speech perception and the brain, 358–362, 359–362 (figure) effect of vision on, 351–352, 351–352 (figure) hearing loss and, 362–364 human voice as stimulus and, 342–347, 343–345 (figure), 346 (table) introduction to, 341–342 theories of, 355–358, 356 (figure) top-down processing and, 352–353, 353 (figure) vowels and consonants in, 343–344, 344 (figure) Speed of sound, 291–292, 292 (figure) Spinothalamic pathway, 413 Sports and stereopsis, 218–219, 219 (figure) Spotlight model of attention, 261, 262 (figure) Stapedius, 301 Stapes, 299 Stars, 31, 31 (figure) Stereocilia, 304 Stereograms, 207–210 (figure), 207–211 Stereopsis, 200–201, 201–202 (figure) developmental issues in, 210–211 sports and, 218–219, 219 (figure) Steven’s power law, 35, 35 (figure), 35 (table) Stimulus, 5 attention-drawing, 264–267, 265–266 (figure) human voice as, 342–347, 343–345 (figure), 346 (table) salience of, 265–266 sound as, 290–297, 290–297 (figure) Stimulus onset asynchrony, 260 Substantia gelatinosa, 415 Subtractive color mixing, 158, 158 (figure) Superior colliculus, 98–99, 98 (figure) Superior olive, 319 Supertasters, 452 Supporting cells, 438 Symmetry, 131, 131 (figure) Synesthesia, 385–386, 386 (figure) Tactile agnosia, 421–422 Tapetum, 82 Target range, 333 Tastants, 447 Taste, 7, 447–448, 447–448 (figure) anosmia and, 440, 453–455 artificial sweeteners and perception of, 455–457, 456 (figure) of chili peppers, 452–453, 452 (figure) coding and anatomy of the tongue, 448–450, 449–450 (figure) flavor and, 451, 451 (figure) perception, individual differences in, 451–452, 452 (figure) perception development, 453, 453 (figure) Taste buds, 449 Taste receptor cells, 449 Tasters, 451, 452 (figure) Tectorial membrane, 304 Tempo, 377–378 Temporal code theory, 305 Temporal segregation, 328, 328 (figure) Tensor tympani, 300–301 Sensation and Perception506 Texting while driving, 256 (figure), 282–283 Texture gradients, 194, 194 (figure) Thermoreception, 407–409, 408 (figure) Thermoreceptors, 408 Timbre, 378–379, 379 (figure) waveforms and, 295–297, 296 (figure) Time attention over, 268–272, 269–271 (figure) to collision, 22 reverberation, 335 Tinnitus, 308, 308 (figure) Tip-of-the-nose phenomenon, 444 Tones, pure, 292, 292 (figure) Tongue, 448–450, 449–450 (figure) Tonotopic organization, 320 Top-down processing, 126–128, 127–128 (figure), 352–353, 353 (figure) Topographic agnosia, 140 Touch and pain haptic perception, 419–420 (figure), 419–422 introduction to, 401–403, 402 (figure) mechanoreception, 404–406, 405 (figure) neural pathways, 410–416, 411–416 (figure) neurochemistry of, 417 nociception and perception of, 409–410, 410 (figure) perception of itch and, 418, 418 (figure) phantom limbs and, 427–428 proprioception, 406–407, 407 (figure) skin and its receptors and, 403–410, 403 (figure), 405 (figure), 407–410 (figure) somatosensory cortex, 413–415, 413–415 (figure) thermoreception, 407–409, 408 (figure) vestibular system and perception of balance, 423–425, 424 (figure) Trachea, 343, 343 (figure) Transduction, 6, 66–67 Transmagnetic stimulation (TMS), 43, 43 (figure) Transposition, 381 Trapezoid body, 319 Traveling wave, 303 Triangle, 135, 135 (figure) Trichromatic theory of color vision, 162–163, 163 (figure) Trigeminal nerve, 440 Tritanopia, 173 Tritone paradox, 394, 394 (figure) Tuberous receptors, 426 Tufted cells, 441 Turbinates, 438 Two-point touch threshold, 31–32 Tympanic canal, 302 Tympanic membrane, 298 Unconscious inference, 11–12 Uncrossed disparity, 204 Unilateral dichromacy, 174 Unilateral visual neglect, 276–277, 277 (figure) Unique colors, 166 Univariance, 161–162 Unvoiced consonant, 344 U.S.S. Vincennes, 37–39, 37 (figure) Uvula, 343, 343 (figure) Van Beethoven, Ludwig, 394–395, 395 (figure) Vanilla, 443–444, 444 (figure) Vehicle collisions, 21–22 Ventral pathway, 107, 109–110, 109 (figure) object perception in, 137–142, 137–142 (figure) See also Brain, visual Ventral posterior nucleus of the thalamus, 413 Vergence, 199–200, 200 (figure), 235–236 Veridicality, 6 Vestibular canal, 302 Vestibular complex, 425 Vestibular system, 423–425, 424 (figure) Viewpoint invariance, 136 Virtual reality and therapy, 219–221, 220–221 (figure) Virtual reality exposure therapy (VRET), 221 Visible spectrum, 153 Vision, neuroscience of motion and, 232–240, 233–235 (figure), 237–240 (figure) Vision prostheses, 85–86, 86 (figure) Visual angle, 212, 212 (figure) Visual brain, 91–92 anatomy and physiology of attention and, 272–277, 273–277 (figure) blindsight and, 113–116, 115 (figure) conjugate gaze palsy and, 116–117 damage and neuropsychology, 18 development of vision and, 111–113 executive attention network, 273–274, 274 (figure) functional pathways in visual cortex of, 107–110, 108–109 (figure) how attention affects, 274–275, 275 (figure) lateral geniculate nucleus, 94–98, 95–97 (figure) lobes, 100, 100 (figure) mapping the eye on the, 101–102, 101–102 (figure) MT as movement area of, 236–239, 237–239 (figure) object perception in, 137–142, 137–142 (figure) optic nerve and chiasm, 92–94, 93 (figure) orienting attention network, 273, 273 (figure) primary visual cortex, 99–106 superior colliculus, 98–99, 98 (figure) V2 and beyond, 106–107, 106 (figure) where vision comes together in, 111 Visual capture, 428 Visual consciousness and awareness, 279–282, 279–282 (figure) Visual detection, 38 Visual illusions of size and depth, 213–217 Visually guided eye movements, 244–245, 245 (figure) Visually guided grasping, 245–246 Visual search, 266–267, 267 (figure), 268 (figure) Visual spectrum, 153 Visual system, 7, 53 development of, 111–113 duplex theory of vision and, 68–73 light and, 55–57 refractive errors and diseases of the eye, 77–81 retinal ganglion cells and receptive fields, 73–75 (figure), 73–77, 77 (figure) role of the eye in, 57–62 as source of joy, 54, 54 (figure) tests of, 45–49 vision prostheses for, 85–86, 86 (figure) See also Eye(s) Voice area, 361, 361 (figure) Voice box, 343, 343 (figure) Voiced consonant, 344 Voicing, 344 Voicing-onset time, 349–350, 350 (figure) Von Békésy, Georg, 305, 306 (figure) Vowels and consonants, 343–344, 344 (figure) V1 region, 101–102, 101–102 (figure) belt and parabelt regions, 321 complex cells, 103–104, 103–104 (figure) opponent cells in, 166–167, 167 (figure) organization of, 104–106, 105 (figure) receptive fields, 102 simple cells, 103, 103 (figure) V2 region, 106–107, 106 (figure) V4 region, 137 attention and, 275, 275 (figure) cortical achromatopsia and, 174–175 V5 region, 110 Warm fibers, 408–409 Waterfall illusion, 10 (figure), 239–240, 240 (figure) Waveform and timbre, 295–297, 296 (figure) Wavelength, 55–56, 56 (figure) of light and color, 153–155, 154 (figure) Waves, sound, 55, 291, 291 (figure) traveling, 303 Weber, Ernst Heinrich, 12, 31 Weber’s law, 12–13 Wernicke, Carl, 359, 359 (figure) Wernicke’s aphasia, 360 Wernicke’s area, 359–360, 359 (figure) Wertheimer, Max, 128, 132–133 Wiesel, Torsten, 91–92, 92 (figure) Windpipe, 343, 343 (figure) Word segmentation, 353 Wundt, Wilhelm, 13 Young, Thomas, 11, 162–163, 163 (figure) Zero disparity, 204 Zonule fibers, 61 SENSATION & PERCEPTION- FRONT COVER SENSATION & PERCEPTION COPYRIGHT BRIEF CONTENTS DETAILED CONTENTS ISLE ACTIVITIES PREFACE ACKNOWLEDGMENTS ABOUT THE AUTHORS CHAPTER 1- WHAT IS PERCEPTION? CHAPTER 2- RESEARCH METHODOLOGY CHAPTER 3- VISUAL SYSTEM: THE EYE CHAPTER 4- VISUAL SYSTEM: THE BRAIN CHAPTER 5- OBJECT PERCEPTION CHAPTER 6- COLOR PERCEPTION CHAPTER 7- DEPTH AND SIZE PERCEPTION CHAPTER 8- MOVEMENT AND ACTION CHAPTER 9- VISUAL ATTENTION CHAPTER 10- THE AUDITORY SYSTEM CHAPTER 11- THE AUDITORY BRAIN AND SOUND LOCALIZATION CHAPTER 12- SPEECH PERCEPTION CHAPTER 13- MUSIC PERCEPTION CHAPTER 14- TOUCH AND PAIN CHAPTER 15- OLFACTION AND TASTE GLOSSARY REFERENCES AUTHOR INDEX SUBJECT INDEX

Order your essay today and save 25% with the discount code: STUDYSAVE

Order a unique copy of this paper

600 words
We'll send you the first draft for approval by September 11, 2018 at 10:52 AM
Total price:
$26
Top Academic Writers Ready to Help
with Your Research Proposal

Order your essay today and save 25% with the discount code GREEN