Learning to See

The most brilliant scientific insight depends, like the everyday faculty of sight, on distinguishing meaningful signals from among random ones.

Anthropology

Current Issue

This Article From Issue

July-August 2018

Volume 106, Number 4
Page 242

DOI: 10.1511/2018.106.4.242

When Galileo looked at the Moon through his new telescope in early 1610, he immediately grasped that the shifting patterns of light and dark were caused by the changing angle of the Sun’s rays on a rough surface. He described something akin to mountain ranges “ablaze with the splendor of his beams,” their sides in shadow like “the hollows of the Earth”; he also rendered these observations in a series of masterful drawings.

Six months before, the English astronomer Thomas Harriot had also turned the viewfinder of his telescope towards the Moon. But where Galileo saw a new world to explore, Harriot’s sketch from July 1609 suggests that he saw a dimpled cow pie. Why was Galileo’s mind so receptive to what lay before his eyes, while Harriot’s vision deserves its mere footnote in history?

Learning to see is not an innate gift; it is an iterative process, always in flux and constituted by the culture in which we find ourselves and the tools we have to hand. Harriot’s 6-power telescope certainly didn’t provide him with the level of detail of Galileo’s 20-power. Yet the historian Samuel Y. Edgerton has argued that Harriot’s initial (and literal) lack of vision had more to do with his ignorance of chiaroscuro—a technique from the visual arts first brought to full development by Italian artists in the late 15th century.

Wellcome Images/Science Source; Max Alexander/Lord Egremont/Science Source

Ad Right

By Galileo’s time, the Florentines were masters of perspective, using shapes and shadings on a two-dimensional canvas to evoke three-dimensional bodies in space. Galileo was a friend of artists, and someone who in his youth might have considered becoming one himself. He believed with a kind of religious fervor that the creator of the world was a geometer. Galileo likely imbibed these mathematically deep methods of representation, based as they are on the projective geometries of light rays. Harriot, on the other hand, lived in England, where general knowledge of these representational techniques hadn’t yet arrived. The first book on the mathematics of perspective in English—The Art of Shadows, by John Wells—appeared only in 1635. When Galileo looked at the face of the Moon, he had no trouble understanding that there, mountaintops first catch fire with the rising Sun while their lower slopes remain in darkness, just as they do on Earth. Galileo, therefore, had a theory for what he was seeing when those pinpricks of light winked into existence along the terminator line of day and night; he even used the effect to measure the heights of those mountains, finding them higher than the Alps. Harriot, a brilliant polymath yet possibly blind to this geometry, looked at the same scenes half a year before Galileo, but didn’t understand.

When we consider scientific observations—those paragons of a purportedly objective gaze—we find, in fact, that they are often complex, contingent, and distributed phenomena, much like human vision itself. Assemblies of high-powered machines that detect the otherwise undetectable, from gravitational waves in the remotest cosmos to the minute signals produced by spinning nuclei within human cells, rely on many forms of “sight” that are neither simple nor unitary. By exploring vision as a metaphor for scientific observation, and scientific observation as a kind of seeing, we might ask: How does prior knowledge about the world affect what we observe? If prior patterns are essential for making sense of things, how can we avoid falling into well-worn channels of perception? And most importantly, how can we learn to see in genuinely new ways?

Finding Orderly Beauty in the Sky

Scientific objectivity is the achievement of a shared perspective. It requires what the historian of science Lorraine Daston and her colleagues call idealization: the creation of some simplified essence or model of what is to be seen, such as the dendrite in neuroscience, the leaf of a species of plant in botany, or the tuning-fork diagram of galaxies in astronomy. Even today, scientific textbooks often use drawings rather than photographs to illustrate categories for students, because individual examples are almost always idiosyncratic; too large, or too small, or not of a typical coloration. The world is profligate in its variability, and the development of stable scientific categories requires much of that visual richness to be simplified and tamed.

In 1890, the meteorologists Hugo Hildebrandsson, Wladimir Köppen, and Georg von Neumayer published a most unusual attempt at idealization. Theirs was the first cloud atlas, proposing to standardize terminologies and categories of wisps of water vapor. To earlier generations, this had seemed a hopeless project. Clouds are things of infinite variety and shape, a drifting canvas for daydreams. But the creation of agreed-upon names and categories—the feathery high-altitude cirrus, the brooding low-lying stratus, the puff-ball cumulus—this parsing of the visual world overhead and the creation of a shared vocabulary was a great advance of 19th-century meteorology.

From International Cloud Atlas by Hugo Hildebrand Hildebrandsson. Courtesy of UC San Diego Library.

Clouds, as it turned out, are a helpful proxy; they allow large numbers of semiskilled observers on the ground to visualize the flow conditions overhead. Following the simple instructions provided by Hildebrandsson and his colleagues, which eventually became an international scientific effort, the observers would know how to report observations of cloud types and the directions of motion, along with local weather conditions. Taken over the course of a year, commencing in May 1896, this compendium of information would allow meteorologists to puzzle out the pattern of winds in the upper atmosphere for the first time.

I was interested in seeing the International Cloud Atlas (1896) first-hand. Even though I am a theoretical physicist, I still harbor a need to hold something in my hand, and to see it with my own eyes, before I believe I’ve grasped its nature. Although the Atlas has been updated many times, and the current version is available online, there are not many copies of the 1896 Atlas left, so I had to take a road trip to visit the Special Collections Library at the University of Virginia. Instead of the book I expected, the librarian presented me with what looked like an artist’s portfolio, bound by a ribbon: eight pages of text and a collection of 28 photochromotypes that were as fragile as old family photographs.

The need to supplement the photographs with drawings and paintings is an example of idealization at work.

The foregrounds are particularly intriguing as attempts to frame and contextualize forms that would otherwise be free-floating. A water tank pokes into the sky in the photograph of the cirrus; rooflines in Paris are photographed beneath mottled stratocumulus; a young man dawdles on the bank of a river, in pastel, gazing up at low-lying stratus; and a sailing ship is painted upon a calm ocean, eternally under way beneath a hazy bank of altostratus. The need to supplement the photographs with drawings and paintings is an example of idealization at work. Photographers had only recently learned how to capture the sky in a manner that brought out the shapes of clouds, and so the first Atlas is a mix of photographs and artful renderings in paint and pastel by human hands. The creation of objective categories of cloud-shapes thus involved a collaboration between art and science, driven by a love of shape and form, and the uncovering of an orderly beauty in the sky overhead.

In our urge to find patterns, we are like a rock climber pulling herself up the sheer wall of the world. She drives a piton in when she finds a handhold of pattern, a small crevice of meaning, some slight imperfection in the rock face. As a physicist, I believe that this works only because there is already some opening there to grab on to. To say that we construct idealized categories is not to say that patterns in the world don’t already exist, but that we must learn how to see them in the world around us.

The Patterns We Need to See

Hold up your hand in front of your face: How can you see what’s there? Parsing meaning from randomness—the signal from the noise—is fundamental to both sight and scientific observation. Unless we are blind, our open eyes are flooded with photons at every moment, a noisy stream of information that is then launched from the retina, traveling as electrochemical impulses along the optic-nerve pathways. These are taken up by neural assemblies and, in the dark cavern of the skull (filled to the brim as it is with a sloshing assembly of soft matter), the brain sifts that welter of data for signals that conform to particular patterns. (Fingers? Check. Five of them? Check.) Some neural assemblies specialize in detecting certain shapes, such as edges or corners; others specialize in collecting those shapes into higher-order schemes, such as a coffee cup, the face of a friend, or your hand.

These internal visual elements are a mix of predilections that we are born with and patterns learned from personal experience; how they affect our perception varies according to our understanding and expectations. When the midcentury psychologists Jerome S. Bruner and Leo Postman presented test subjects with brief views of playing cards, including some nonstandard varieties—such as a red two of spades, or a black ace of diamonds—many people never called out the incongruities. They reported that they felt uneasy for some reason but often couldn’t identify why, even though the reason was literally right before their eyes.

Illustration by Barbara Aulicino.

So, crucially, some understanding of the expected signal usually exists prior to its detection: To be able to see, we must know what it is we’re looking for, and predict its appearance, which in turn influences the visual experience itself. The process of perception is thus a bit like a Cubist painting, a jumble of personal visual archetypes that the brain enlists from moment to moment to anticipate what our eyes are presenting to us, thereby elaborating a sort of visual theory. Without these patterns we are lost, adrift on a sea of chaos, with a deeply unsettling sense that we don’t know what we are looking at, yet with them we risk seeing only the familiar. How do we learn to see something that is truly new and unexpected?

Vision is not only personal and patterned, but also complex and spatially distributed. In the 1970s, Elizabeth K. Warrington and Angela M. Taylor studied patients with brain damage in the posterior regions of the brain, but with no apparent damage to their visual pathways. It transpired that the part of the brain that’s active when we identify the three- dimensional shape of an object (say, a cylindrical white item on the desk) is different from the area involved in knowing its purpose or name (a coffee cup that holds your next sip). Warrington and Taylor tested subjects’ performance on a “same shape” task by presenting photographs of familiar objects, such as a telephone, but taken from both standard and nonstandard perspectives.

Portrait of Pablo Picasso, by Juan Gris, 1912. The Art Institute of Chicago, gift of Leigh B. Block, 1958.525.

For the “same function” task, all the objects were photographed in a standard view, but subjects had to group them according to function—for example, to distinguish different sorts of telephones from a mailbox. People with lesions in the right posterior region showed no deficit when objects were presented “normally,” but displayed a reduced ability to identify objects when they were seen from unusual perspectives. Roughly speaking, the results were consistent with the subjects being unable to rotate the object in the mind’s eye into a more “standard” orientation for comparison with internal idealized mental models. Meanwhile, people with lesions in the left posterior brain showed little deficit in such geometric categorization, yet displayed a reduced ability to identify the object’s name or function.

Subsequent research, now including functional MRI imaging on healthy patients, has led to more nuanced understandings of these matters, but the key insight remains: Our extraction of a visual signal is a process distributed across the brain, a silent chorus of the mind that—when it works—produces effortless extraction of visual meaning.

If the brain is a taxonomizing engine, anxious to map the things and people we experience into familiar categories, then true learning must always be disorienting. Learning shifts the internal constellation of the firings of our nerves, the star by which we set our course, the spark of thought itself. This mental flexibility is an undeserved inheritance, hard-won over eons by our ancestors, and it serves as a good metaphor for how scientists can learn to see with new machine eyes.

Perceiving Invisible Patterns

In science, seeing things afresh sometimes demands a concerted (and contested) shift in paradigms, such as the move from Ptolemy’s map of the planets to those of Copernicus and Galileo. On other occasions, it happens by accident. In a fundamental sense, all of the output of our instruments is signal; noise is just that part we are not interested in. This view means that separating out the signal depends upon who is doing the observing and for what purpose. The “noise” picked up by Arno Penzias and Robert Wilson in 1964 in their microwave detectors at Bell Labs in New Jersey turned out to be a clue to one of the most astonishing discoveries in science. After heroic efforts to reduce the background hiss detected by their instrument, they determined that it was not coming from the instrument itself, nor from its suburban surroundings. Instead, they convinced themselves that it was omnidirectional, and coming from the sky. They eventually realized that the hiss was part of the cosmic background, remnant radiation and a “signal” from the Big Bang. Now, 50 years later, the detailed angular distribution across the sky of statistical fluctuations in that radiation is a key piece of evidence that leads many cosmologists to believe in dark energy.

From The Trouvelot Astronomical Drawings Manual, by E. L. Trouvelot (1882). Courtesy of the New York Public Library.

The scientific projects that best stretch our understanding of what it means to see in new ways, and reveal the distributed nature of observation, are those that span the globe and involve thousands of individuals, such as the Laser Interferometer Gravitational-Wave Observatory (LIGO). Efforts such as LIGO are often described as “opening a new window” on the cosmos—yet no human being sees a gravitational wave with LIGO, just as no single neuron in our brain sees a bluebell flower. In both cases, seeing is a multipartite process, requiring a comparison between noisy signals and idealized models.

More than a century ago, Albert Einstein predicted that a warpage of space and time propagates at the speed of light, like the outgoing ripple from the snap of a rug. These space-time ripples are gravitational waves. Consider an incoming burst of gravitational waves, emitted from a distant galaxy by a colliding pair of black holes. Gravity is extraordinarily weak compared with the electromagnetic forces that hold most objects together in our world, giving everyday stuff a solidity that we take for granted. The weakness of gravitational waves therefore means that they have completely negligible effects locally, unless we isolate massive objects from one another and then allow them to move freely in response to the passing space-time ripples. The gravitational waves generate oscillations in the relative positions of these massive objects, not unlike the motions of leaves afloat on the surface of a pond as a ripple passes among them.

To say that we construct idealized categories is not to say that patterns in the world don’t already exist, but that we must learn how to see them in the world around us.

If Einstein was right, we are bathed in these ripples, but we have been blind to the information they carry—until now. In the case of LIGO, the passage of a gravitational wave produces a shudder—a fraction of a nuclear diameter—in the relative positions of two pairs of 40-kilogram mirrors, four kilometers apart. Splitting a specially prepared laser beam in two, bouncing it off these mirrors, and allowing the two laser beams to interfere again, the LIGO detector measures a time-varying interference pattern that can be fit to theoretical predictions.

In tandem with the measurement, thousands of simulations are run using the Einstein field equations, the mathematical language of general relativity. These simulations generate candidate signals—essentially a kind of “gravitational wave atlas”a numerical compendium of idealizations of what LIGO’s output would look like if certain events had occurred in the distant cosmos. These simulations are answers to “What if?” questions such as: What if two black holes, each 10 solar masses in size, collided 5 billion light-years away? What if one of the black holes was 5 solar masses in size and the other 15, and the distance was 3 billion light-years? By varying the masses and orbital characteristics that initialize the simulation, along with the distance to the event, a best fit to the measured signal is found. If the fit between the measured and idealized model signals is good enough, the LIGO team then infers that two black holes likely collided a billion light-years away. And so, humanity can now see space-time ripples where before we were gravitational-wave blind.

If creating a new telescope, from radio to the gamma-ray, is like adding another electromagnetic eye, LIGO is really like adding an entirely new sense organ. What’s more, our new sense organ can see directly into regions of the universe that were opaque to all of our previous electromagnetic eyes.

The Value of Prior Knowledge

Because of the complexity of both visual experience and scientific observation, it is clear that although seeing might be believing, it is also true that believing affects our understanding of what we see. The filter we bring to sensory experience is commonly known as cognitive bias, but in the context of a scientific observation it is called prior knowledge. To call it prior knowledge does not imply that we are certain it is true, only that we assume it is true in order to get to work making predictions. For example, LIGO researchers assume that the Einstein field equations are universal. This implies that the equations can be tested locally, but they are also assumed to govern the motion of orbiting black holes a billion light-years away.

If we make no prior assumptions, then we have no ground to stand on. The quicksand of radical and unbounded doubt opens beneath our feet and we sink, unable to gain purchase. We remain forever at the base of the sheer rock face of the world, unable to begin our climb. Yet, although we must start with prior knowledge we take as true, we must also remain open to surprise; else we can never learn anything new. In this sense, science is always Janus-headed, like the ancient Roman god of liminal spaces, looking simultaneously to the past and to the future. Learning is essentially about updating our biases, not eliminating them. We always need them to get started, but we also need them to be open to change, otherwise we would be unable to exploit the new vistas that our advancing technology opens to view.

If the brain is a taxonomizing engine, anxious to map the things and people we experience into familiar categories, then true learning must always be disorienting.

The invention of the telescope heralded a new era of human sight. It led to a flurry of new observations—and great puzzlement. Astronomers were repeatedly confronted with images in their viewfinders that they struggled to make sense of, not only in terms of finding a physical theory for what they were seeing, but something more basic: They struggled to identify stable and repeatable visual patterns, to create a kind of astronomical equivalent to the Cloud Atlas.

In Planets and Perception (1988), the astronomy historian William Sheehan notes that a good way to reveal that we don’t understand something is to attempt to draw it. Hence his interest in the early sketches by telescopic observers of the Moon, Mars, and the rings of Saturn. What aspects in the visual field were the most essential? What pieces among the collection of edges, shapes, and textures connected to one another in sensible structures? What was closer, and what was farther away? What was shadowing, and what was coloration? Until Galileo arrived to answer many of these questions, astronomers struggled to interpret their nightly visual experience, and these are the same problems faced today by researchers who are trying to create machine-vision systems.

The Unchartable Frontiers of Science

The iterative bootstrapping of learning-to-see, then seeing-to-learn, continues apace. But in the four centuries since Galileo bent to look through his glazed optic tube, the human brain has not changed all that much, if at all. Rather, the current revolution comes from our new tools, new theories, and new methods of analysis made possible by new hardware. Detectors make visible what was previously hidden, and the learning-to-see half of the feedback loop involves ever more powerful and subtle computer algorithms that seek patterns in those new observations. As Daston argues, the objectivity of scientific observation means parsing the world into pieces, and naming those pieces through shared idealizations. But this is now done using a data stream from a global network of detectors and telescopes, aided by smart algorithms to assist in our naming, learning to navigate an information flood that each second dwarfs the amount of data collected by Galileo in a lifetime of observations. Our machines have given us new eyes so we can see things in the world that have been there all along.

Statistical Atlas of the United States, Based Upon the Results of the Eleventh Census (1890).

If there are frontiers in science, admittedly a fraught metaphor, then we will forever be in the state of the American frontier at the 1890 census—with areas thought roughly settled that are in fact broken up into isolated pockets with immense gaps in-between. The evolutionary biologist E. O. Wilson argues that there are likely still millions of species undiscovered. Where are they? Underfoot, undersea, and lurking in still-remote places. We have not plumbed the depths of the natural world, even on our familiar Earth.

If we cast our thoughts outward to the wider Solar System and beyond, the mind boggles at how much there is to learn. The world is infinite in all directions, as the theoretical physicist Freeman Dyson wrote in 1988. Outward to the stars, inward to the nucleus, and, casting sideways, we find the infinite complexity of the biosphere and human cultures, the crenellations and foldings of the human neocortex that somehow contains so much that is light and dark in our beings. An army of scientists will never be enough to grasp the whole of it. That is an open-ended project for the species. Let’s hope that we are always like Galileo setting up his telescope for a night’s viewing, prepared to be astonished, ready to see in new ways, our minds like tinder awaiting a spark.


This article has been adapted from one that was published in Aeon.

Bibliography

  • Bruner, J. S. and L. Postman. 1949. On the perception of incongruity: A paradigm. Journal of Personality 18:206–223.
  • Daston, L. and P. Galison. 1992. The image of objectivity. Representations 40:81–128.
    • Sheehan, W. 1988. Planets and Perception: Telescopic Views and Interpretations, 16091909. Tucson, Arizona: University of Arizona Press.
    • Warrington, E. K. and A. M. Taylor. 1978. Two categorical stages of object recognition. Perception 7:695–705.

American Scientist Comments and Discussion

To discuss our articles or comment on them, please share them and tag American Scientist on social media platforms. Here are links to our profiles on Twitter, Facebook, and LinkedIn.

If we re-share your post, we will moderate comments/discussion following our comments policy.