Jump to content

Max Mathews

From Wikipedia, the free encyclopedia
(Redirected from RTSKED)
Max Mathews on his 80th birthday

Max Vernon Mathews (November 13, 1926 in Columbus, Nebraska, US – April 21, 2011 in San Francisco, CA, US) was an American pioneer of computer music.

Biography

[edit]
Max Mathews playing one of the electronic violins he built, in his analog electronics lab at Bell Telephone Labs (c. 1970)

Max Vernon Mathews was born in Columbus, Nebraska, to two science schoolteachers. His father in particular taught physics, chemistry and biology in the Peru High School [1] of Nebraska, where he was also the principal.[2] His father allowed him to learn and play in the physics, biology and chemistry laboratories, where he enjoyed making lots of things from motors to mercury barometers. At the age of 9, when students are usually introduced to algebra, he started to study by himself the subject with few other students. That was because the vast majority of population there were farmers and their sons weren't interested about learning algebra, since it isn't useful for the everyday work. In the same way he studied calculus, but he never graduated from high school.[2]

After a period as a radar repairman in the navy, where he fell in love with electronics, Mathews decided to study electrical engineering at the California Institute of Technology and the Massachusetts Institute of Technology, receiving a Sc.D. in 1954. Working at Bell Labs, Mathews wrote MUSIC, the first widely used program for sound generation, in 1957. For the rest of the century, he continued as a leader in digital audio research, synthesis, and human-computer interaction as it pertains to music performance. In 1968, Mathews and L. Rosler developed Graphic 1, an interactive graphical sound system on which one could draw figures using a light-pen that would be converted into sound, simplifying the process of composing computer generated music.[3][4] Also in 1970, Mathews and F. R. Moore developed the GROOVE (Generated Real-time Output Operations on Voltage-controlled Equipment) system,[5] a first fully developed music synthesis system for interactive composition and realtime performance, using 3C/Honeywell DDP-24[6] (or DDP-224)[7] minicomputers. It used a CRT display to simplify the management of music synthesis in realtime, 12bit D/A for realtime sound playback, an interface for analog devices, and even several controllers including a musical keyboard, knobs, and rotating joysticks to capture realtime performance.[3][7][4]

Although MUSIC was not the first attempt to generate sound with a computer (an Australian CSIRAC computer played tunes as early as 1951),[8] Mathews fathered generations of digital music tools. He described his work in parental terms, in the following excerpt from "Horizons in Computer Music", March 8–9, 1997, Indiana University:

Computer performance of music was born in 1957 when an IBM 704 in NYC played a 17 second composition on the Music I program which I wrote. The timbres and notes were not inspiring, but the technical breakthrough is still reverberating. Music I led me to Music II through V. A host of others wrote Music 10, Music 360, Music 15, Csound and Cmix. Many exciting pieces are now performed digitally. The IBM 704 and its siblings were strictly studio machines – they were far too slow to synthesize music in real-time. Chowning's FM algorithms and the advent of fast, inexpensive, digital chips made real-time possible, and equally important, made it affordable. Starting with the GROOVE program in 1970, my interests have focused on live performance and what a computer can do to aid a performer. I made a controller, the Radio-Baton, plus a program, the Conductor program, to provide new ways for interpreting and performing traditional scores. In addition to contemporary composers, these proved attractive to soloists as a way of playing orchestral accompaniments. Singers often prefer to play their own accompaniments. Recently I have added improvisational options which make it easy to write compositional algorithms. These can involve precomposed sequences, random functions, and live performance gestures. The algorithms are written in the C language. We have taught a course in this area to Stanford undergraduates for two years. To our happy surprise, the students liked learning and using C. Primarily I believe it gives them a feeling of complete power to command the computer to do anything it is capable of doing.

Radio-Baton demonstration by Mathews on SF MusicTech Summit 2010

In 1961, Mathews arranged the accompaniment of the song "Daisy Bell" for an uncanny performance by computer-synthesized human voice, using technology developed by John Kelly, Carol Lochbaum, Joan Miller and Lou Gerstman of Bell Laboratories. Author Arthur C. Clarke was coincidentally visiting friend and colleague John Pierce at the Bell Labs Murray Hill facility at the time of this remarkable speech synthesis demonstration and was so impressed that he later told Stanley Kubrick to use it in 2001: A Space Odyssey, in the climactic scene where the HAL 9000 computer sings while his cognitive functions are disabled.[9]

Mathews directed the Acoustical and Behavioral Research Center at Bell Laboratories from 1962 to 1985, which carried out research in speech communication, visual communication, human memory and learning, programmed instruction, analysis of subjective opinions, physical acoustics, and industrial robotics. From 1974 to 1980 he was the Scientific Advisor to the Institute de Recherche et Coordination Acoustique/Musique (IRCAM), Paris, France, and from 1987 was Professor of Music (Research) at Stanford University. He served as the Master of Ceremonies for the concert program of NIME-01, the inaugural conference on New interfaces for musical expression.

Mathews was a member of the National Academy of Sciences, the National Academy of Engineering and a fellow in the American Academy of Arts and Sciences, the Acoustical Society of America, the IEEE, and the Audio Engineering Society. He received a Silver Medal in Musical Acoustics[10] from the Acoustical Society of America, and the Chevalier de l'ordre des Arts et Lettres, République Française.

The Max portion of the software package Max/MSP is named after him (the MSP portion is named for Miller Puckette, who teaches at UC San Diego).

Mathews died on the morning of 21 April 2011 in San Francisco, California of complications from pneumonia. He was 84. He was survived by his wife, Marjorie, his three sons and six grandchildren.

See also

[edit]

References

[edit]
  1. ^ Bell, Gordon C. "Max V. Mathews 1926–2011". National Academy of Engineering. Retrieved 16 May 2023.
  2. ^ a b Park, Tae Hong (2009). "An Interview with Max Mathews". Computer Music Journal. 33 (3): 9–22. JSTOR 40301041.
  3. ^ a b Holmes, Thom (2008). "Digital Synthesis and Computer Music". Electronic and experimental music: technology, music, and culture. Taylor & Francis. pp. 254. ISBN 978-0-415-95781-6.
  4. ^ a b Roads, Curtis (Winter 1980). "Interview with Max Mathews". Computer Music Journal. 4 (4).
    in Curtis Roads, ed. (1989). The Music Machine: Selected Readings from Computer Music Journal. MIT Press (1989/1992). pp. 5. ISBN 978-0-262-68078-3.
  5. ^ Max V., Mathews; F.R., Moore (1970). "GROOVE—a program to compose, store, and edit functions of time". Communications of the ACM. 13 (12).
  6. ^ Nyssim Lefford; Eric D. Scheirer & Barry L. Vercoe. "An Interview with Barry Vercoe". Experimental Music Studio 25. Machine Listening Group, MIT Media Laboratory.
  7. ^ a b Bogdanov, Vladimir (2001). All music guide to electronica: the definitive guide to electronic music. Backbeat Books. pp. 320. ISBN 978-0-87930-628-1.
  8. ^ "Australian CSIRAC computer website".
  9. ^ "Bell Labs: Where "HAL" First Spoke (Bell Labs Speech Synthesis Web Site)". Archived from the original on April 7, 2000. Retrieved February 8, 2007.
  10. ^ "Acoustical Society of America Awards". Acoustical Society of America. Retrieved 2021-11-28.
[edit]