Legacy: Decades of Discovery - 1970s

In 1970, Haskins Laboratories completed the move to New Haven, CT, begun in 1969, and entered into affiliation agreements with Yale University and the University of Connecticut.

Recognizing the Laboratories’ unique facilities for analysis and synthesis of speech, the National Institutes of Health defrayed the costs of sharing the facilities with investigators from other institutions — support that continued for nearly twenty years.

Katherine S. Harris, working with Fredericka Bell-Berti, Gloria Borden, and others, demonstrated electromyographically how the precise phasing and layering of articulatory actions give rise to segmental overlap, and thus to the acoustic phenomena of coarticulation.

Tim Rand discovers “duplex perception.”

Isabelle Liberman, Donald Shankweiler, and Alvin Liberman teamed up with Ignatius Mattingly to study the relation between speech perception and reading, a topic implicit in the Laboratories’ research program since the 1940s. They developed the concept of “phonemic awareness,” the knowledge that would-be readers must have of the phonemic structure of their language if they are to learn to read. Under the broad rubric of the “Alphabetic Principle,” this concept is the core of the Laboratories’ program of reading pedagogy today.

Patrick Nye joined the Laboratories to lead a team including Franklin Cooper, Jane Gaitenby, George Sholes, and Gary Kuhn in work on the reading machine. The project culminated when the addition of an optical typescript reader enabled investigators to assemble the first automatic text-to-speech reading machine. By the end of the decade the technology had advanced to the point where commercial concerns assumed the task of designing and manufacturing reading machines for the blind.

Working with Bruno Repp, Virginia Mann, Joanne Miller, Douglas Whalen, and others over the next decade or so, Alvin Liberman conducted a series of innovative experiments to clarify and deepen the concept of a speech mode of perception. These experiments moved away from the cue as a static property of the acoustic signal toward the cue as a dynamic index of articulatory action.

Experiments by Peter Bailey, James Cutting, Michael Dorman, Quentin Summerfield, and others cast doubt on the validity of the “acoustic cue” as a unit of perceptual function. Building on these experiments, Philip Rubin developed the sinewave synthesis program used by Robert Remez, Rubin, David Pisoni, and colleagues. These researchers showed that listeners can perceive continuous speech, without traditional speech cues, from a pattern of three sinewaves that track the changing resonances of the vocal tract. Their work paved the way for a view of speech as a dynamic pattern of trajectories through articulatory-acoustic space.

Rubin, Thomas Baer, Paul Mermelstein, and colleagues developed Mermelstein’s anatomically simplified vocal tract model into the first articulatory synthesizer that can be controlled in a physically meaningful way and used for interactive experiments

1930s

1940s

2000s

2010s

2020s

1970s

1980s

1990s