Haile (robot)
Haile (pronounced Hi-lee) is a robot percussionist developed by the Georgia Institute of Technology that listens to music in real time and creates an accompanying beat. The robot was designed in 2006 by Georgia Tech's professor of musical technology, Gil Weinberg. He and one of his graduate students, Scott Driscoll, created the robot to be able to "listen like a human, [and] improvise like a robot" (Weinberg).[1] Haile "listens" through a microphone mounted on the drum and analyzes the sound, separating it into beats, rhythms, pitches and several other qualities. Detecting changes in these qualities helps Haile to assume either a leading or following style of play, roles that define the robot's collaborative abilities. Haile was also the first robot to create an acoustic percussion experience rather than play music through speakers. Its anthropomorphic design, which gives it movable arms that can move in any direction, allow it to create this acoustic music.
Goals and purpose
[edit]Driscoll's initial goal for making Haile was to combine the use of auditory input and robotics to create a musical experience that would lead to further human–robot interaction. The final goal was a robot that could translate live music into an acoustic performance that could implement and transcend human capabilities. Haile wasn't designed to replace human musicians, but rather to accompany them with expressive playing.[2]
These goals led to Weinberg wanting to create an acoustic musical experience. His earlier experiments failed to incorporate the visual or auditory aspects associated with acoustic music. Haile's functional drumming arms add musical cues (visually stimulating bouncing drum sticks and live, acoustic sounds) that other robot performances lack.[1] Additionally, other attempts at percussion playing robots, Weinberg saw, were limited in the variety of beats that they could produce. Haile is not only preloaded with individual beats, but is also programmed to identify pitch, rhythm, and patterns, allowing it to improvise and play different beats every time, rather than simply mimic.[3]
Design
[edit]Haile's anthropomorphic or human-like design mimics human movements which supports interactive play with other musicians. Its two robotic arms are responsible for creating different sounds; the right arm plays faster notes, while the left arm has larger motions for louder and slower beats. While other robotic drummers at the time were limited to playing only a few locations on the drum, Haile is versatile in its ability to play along a straight line from the rim to the center of the drum.[1]
Form
[edit]Haile's wooden designed was modeled to match the natural feel of a Native American pow wow (Native American gathering), so it was made out of wood rather than metal. The wooden parts were made at the Advanced Wood Product Laboratory Archived 2014-11-24 at the Wayback Machine at Georgia Tech's College of Architecture using a CNC wood router. It was originally designed to play a pow wow drum, a multiplayer drum that supports Haile's collaborative purpose. However, it also was made with metal joints that give it adjustable height so it can play other drums. These joints are the keys to the robot's arm movement up and down, left and right, and front and back. If needed, they detach, allowing for full disassembly.[4]
Perception
[edit]Haile uses a microphone on the drum which first detected rhythms played by a human in real time. The robot identifies tempo and beats allowing it to play along with the other player. It can also adjust to the human's changes in volume, tempo or beat, allowing it to switch between accompanying and lead playing.[5] Weinberg and his team first developed the robot's low-level perception ability, which includes detecting hit onsets, pitch, amplitude, and density. In terms of sound, a hit refers to a distinct change in both volume and sound quality. Once the outside music is captured, the sound is analyzed through a number of instruments, called perception modules, which each detect a certain aspect of the sound:
- Pitch – detects hits and changes in frequency and translates them to find pitches[3]
- Beat – processes onsets and determines rhythms and tempo[3]
- Amplitude – recognizes changes in volume to determine when to assume leading or following roles[3]
- Density – detects changes in rhythm complexity at tempo to also help Haile assume leading or following roles[6]
Arm mechanics
[edit]Haile's arms are driven by two separate means. The left arm uses a linear motor which is responsible for larger movements that correspond to louder sounds. The linear motor, along with a linear encoder is used to control the robot's arm height. The larger motions are louder and more visible, but limit the arm to swinging at a top speed of 11 Hz. The right arm, which plays softer and faster notes, is driven by a solenoid, which uses an aluminum stick and return spring capable of hitting the drum at up to 15 Hz. Both arms are capable of hitting anywhere (from the rim to the center) on the drum through the use of a linear slide, which allows each of them completely independent movement in a single direction.
Playing
[edit]Haile's system adopts a leader-follower model, using tempo and beat changes to determine who the current leader is. Haile understands when a new leader emerges based on musical changes (tempo, volume, beat, etc.).[5] The robot has two modes of play:
- As a follower, Haile first analyzes the external music. It then matches and maintains the tempo, allowing the human player to play more complicated rhythms. Haile can also tell when the other player begins to play louder or more quickly, forcing it into the submissive role. When the human beings to play basic rhythms at a steady tempo, the robot takes the lead.[5]
- As a leader, Haile uses rhythms produced earlier by the human and improvises a rhythm with its right arm. The left arm detects and maintains the other player's tempo.[5]
Challenges
[edit]Some of the challenges that Weinberg faced with Haile's programming involved being able to distinguish between different, simultaneous sounds. Initially, analysis algorithms were unable to pick out softer and more subtle notes amidst louder sounds. Also, the inability to filter out ambient noise prevented Haile from working properly. After a considerable amount of adjusting, the filters and input hardware were tuned to differentiate between various volumes of music while ignoring interfering noise.[2]
As Haile was designed to play in either leading or following roles, early detection algorithms limited the human's ability to lead. The robot was designed to be able to detect changes in the music it heard, but would only respond to changes in tempo. This flaw only allowed the human to lead as long as he or she kept speeding up or slowing down. Weinberg, trying to model human musical interaction, implemented volume and noise density sensors to aid in the robot's ability to define leadership. These additions gave the human longer periods of leadership, giving Haile more opportunity to build upon what it heard.[5]
References
[edit]- ^ a b c Weinberg, Gil; Driscoll, Scott (2007). "The design of a robotic marimba player". Proceedings of the 7th international conference on New interfaces for musical expression – NIME '07. p. 228. doi:10.1145/1279740.1279786. ISBN 978-1-4503-7837-6. S2CID 6727904.
- ^ a b Weinberg, Gil; Driscoll, Scott; Parry, Mitchell (2005). "Haile-An Interactive Robotic Percussionist" (PDF). International Computer Music Conference Proceedings. 2005. hdl:2027/spo.bbp2372.2005.069. Archived from the original (PDF) on 2016-03-05. Retrieved November 8, 2014.
- ^ a b c d Abshire, Matthew. "Musical robot composes, performs, and teaches". CNN.com. Retrieved October 26, 2014.
- ^ Weinberg, Gil; Driscoll, Scott (22 April 2006). "Robot-human interaction with an anthropomorphic percussionist". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. pp. 1229–1232. doi:10.1145/1124772.1124957. ISBN 978-1-59593-372-0. S2CID 10301135.
- ^ a b c d e Weinberg, Gil; Blosser, Brian (9 March 2009). "A leader-follower turn-taking model incorporating beat detection in musical human-robot interaction". Proceedings of the 4th ACM/IEEE international conference on Human robot interaction. pp. 227–228. doi:10.1145/1514095.1514149. ISBN 978-1-60558-404-1. S2CID 1200705.
- ^ Weinberg, Gil; Driscoll, Scott (10 March 2007). "The interactive robotic percussionist: New developments in form, mechanics, perception and interaction design". Proceedings of the ACM/IEEE international conference on Human-robot interaction. pp. 97–104. doi:10.1145/1228716.1228730. ISBN 978-1-59593-617-2. S2CID 130547.