| Sound, of course, has always been vital to poetry. As a language 
                    art, the sonic elements of poetry-- accent and duration, syntax 
                    and line, like and unlike sounds, blank verse and free verse-- 
                    have differentiated poetry from prose for thousands of years. 
                     Although the traditional medium of poetry is the human body, 
                    the emergence of new acoustic technologies like the phonograph, 
                    telephone, microphone, loudspeaker, radio, tape recorder, 
                    and more recently, digital audio and surround sound, have 
                    altered the range, volume, reach, and distance of the human 
                    voice, and prompted new literary experiments that investigate 
                    the qualities, characteristics, and material dimensions of 
                    these sound transmitting and recording technologies. When magnetic tape cassettes and stereo tape recorders were 
                    mass-produced for the first time in the 1960's, new possibilities 
                    were made available for cultural production and representation. 
                    As Katherine Hayles points out, the phonograph produced objects 
                    that could be consumed only in their manufactured form, whereas 
                    magnetic tape allowed the consumer to be a producer as well. 
                     As the technology 
                    became more sophisticated, affordable, and widely available, 
                    tape became a popular medium for electronic artists and musicians 
                    to experiment. These experiments ranged from Stockhausen's 
                    1950's tape pieces to the work of minimalist composers like 
                    Steve Reich. Early in his career, Reich composed two works for tape: "It's 
                    Gonna Rain,"(1964) and "Come Out" (1966), which 
                    introduced the concept of "phasing," a process Reich 
                    developed in which two tape loops begin by playing synchronously, 
                    but slowly move out of phase with each other before coming 
                    back into unison. The result is powerfully hypnotic; words 
                    and sentences are collapsed into short phonemes, the building 
                    blocks of language uttered as repetitive sounds that, after 
                    time, morph into new configurations. In the liner notes, Reich describes the process he employed 
                    and the background that inspired the work: 
                    Composed in 1966, Come Out was originally part of a benefit 
                      presented at Town Hall in New York City for the retrial, 
                      with lawyers of their own choosing, of the six boys arrested 
                      for murder during the Harlem riots of 1964. The voice is 
                      that of Daniel Hamm, now acquitted and then 19, describing 
                      a beating he took in Harlem's 28th Precinct Station. The 
                      police were about to take the boys out to be "cleaned 
                      up" and were only taking those that were visibly bleeding. 
                      Since Hamm had no actual open bleeding he proceeded to squeeze 
                      open a bruise on his leg so that he would be taken to the 
                      hospital. 'I had to like open the bruise up and let some 
                      of the bruise blood come out to show them.' Come Out is 
                      composed of a single loop recorded on both channels. First 
                      the loop is in unison with itself. As it begins to go out 
                      of phase a slowly increasing reverberation is heard. This 
                      gradually passes into a canon or round for two voices, then 
                      four voices and finally eight. A complex interplay is set up between the representational 
                    codes of what is spoken or performed and the specificities 
                    of the transmitting or recording technology. In another tape experiment by Alvin Lucier, "I am Sitting 
                    in a Room," (1970), several sentences of recorded speech 
                    are played back into a room where they are re-recorded multiple 
                    times. Over the course of this process, which goes on for 
                    about fifteen minutes, the resonant frequencies of the space 
                    act as a filter, and Lucier's speech (and speech impediment) 
                    is transformed into pure sound. I mention these early tape works by means of an introduction, 
                    or perhaps, more accurately, as an inspiration for thinking 
                    about sound and Web media. Like the tape recorder, new media 
                    editing software is changing the dynamics of who is able to 
                    produce interactive audio-video materials, and it too offers 
                    a rich site to probe the relationship between the technology 
                    and sound. While artists who worked with analog tape media could employ 
                    techniques like cutting and splicing, looping, tape echo, 
                    and direction and speed changes, digital media artists face 
                    new challenges orchestrating and organizing the seemingly 
                    endless possibilities that editing software makes available. 
                   Unlike analog technology, digital technology can be perfectly 
                    precise, giving rise to new practices and techniques that 
                    were not formerly possible. One such moment of digital triumph 
                    occurred when avant-garde musician and composer Georg Anheil's 
                    masterwork, Ballet mécanique, was performed 
                    for the first time in its original instrumentation 75 years 
                    after it was composed.  When Arnheim wrote the music for Ballet mécanique 
                    in 1924, his production called for three xylophones, four 
                    bass drums, a tamtam (gong), two pianos, a siren, three airplane 
                    propellors, seven electric bells, and 16 synchronized player 
                    pianos (or pianolas as they were called then).  Because it was impossible to perfectly synchronize the player 
                    pianos, the work existed as a conceptual piece. Antheil produced 
                    other versions, but he never heard the original in his lifetime. 
                    It wasn't until 1999 when William Holab and Paul Lehrman hooked 
                    up 16 MIDI-compatible (Musical Instrument Digital Interface, 
                    the standard computer protocol for musical instruments) player 
                    pianos to a central sequencer, which enabled all of them to 
                    play in perfect synchronization.  Of course, the artists today that are exploring the new possibilities 
                    for expression that audio-visual Web technologies make available 
                    are still subject to limitations. Looping is easier than it's 
                    ever been to create, but lengthy compositions are a major 
                    challenge. The mp3 file format greatly improved the audio 
                    quality of Web sound, but the file size is still huge compared 
                    to that of text and images. In addition, adding interactivity to sound work often involves 
                    importing the sound files into another software program like 
                    Flash or Director. Subjecting the work to the rules of another 
                    layer of software programming shapes the possibilities of 
                    the final composition.  The three works featured in this Sound issue of PTG were 
                    all created using Flash software. We might see this as a limitation 
                    imposed by a proprietary technology (as some Flash naysayers 
                    might point out), but it also allows us to see a structured 
                    investigation of the software as medium, and the way these 
                    works are expressed through the sound capabilities of Flash. Sounds in Flash can be controlled by means of "attaching 
                    sounds" to "Sound Objects" using ActionScript 
                    code. (An "Object" is a scripting concept that represents 
                    a collection of data and methods for manipulating that data. 
                    For example, the Date Object stores different pieces of information 
                    that relate to time, as well as methods for getting or setting 
                    different values, like the current time.)  
                   This can be used to create work that allows viewers to manipulate 
                    volume and pan controls by sliding a graphic on the screen, 
                    a technique that Jason Nelson explores in Conversation, 
                    which is organized around three subjects: injuries, robots, 
                    and products. Each section is comprised of a series of volume 
                    and pan sliders against a loud graphical display that allows 
                    the user to shift the voices from left to right channels, 
                    and to turn up and down the volume of the often humorous commentators, 
                    who offer looped fragments of stories that relate to products, 
                    injuries, and robots respectively. As the user manipulates the slider controls, the voices result 
                    in a bubbling crowd of conversation, with each commentary 
                    sewn together so that it becomes difficult to find the beginning 
                    or end of any story. The user becomes a DJ selecting voices 
                    to silence or spotlight in the construction of this "verbal 
                    composition." In soundpoem 
                    2, Joerg Peringer uses the selective repetition of short 
                    words, phonemes and letter combinations to investigate the 
                    relationship between words, sounds, and their absences. In 
                    soundpoem 1, 
                    Peringer applies a similar technique by associating repetitive 
                    sounds with specific spaces within the screen. His polite 
                    directions, "please drag the circles into the squares" 
                    stand in shocking contrast to the resulting cacophony that 
                    is revealed to the user who follows directions. Finally, Neil Jenkin's generative poem-engine Orbital 
                    plays with the ideas of space, location, correspondance, and 
                    anonymity. Domain name servers exist to translate numerical 
                    hard-to-remember IP addresses into the familiarity of words. 
                    In this piece, a droning computer voice endlessly lists the 
                    IP numbers of each visitor to the project, while the generative 
                    text engine runs on two databases-- one that contains the 
                    words to Dunlop's text, and another that lists logged IP addresses 
                    of visitors to the engine. As Jenkins describes it, the engine 
                    is programmed using Perl with Flash's Actionscript to count 
                    back through the IP address list and plot the next word in 
                    the poem in a three dimensional plane, using the first three 
                    numbers of the IP address as its x, y and z co-ordinates. 
                    The fourth number in the IP address determines the next word 
                    from the poem to be displayed. Enjoy these works, and as always, feel free to add your comments 
                    to the discussion 
                    board.  M. 
                    Sapnar ..... (1) See Robert Pinsky's The Sounds 
                    of Poetry: A Brief Guide  (Farrar, Straus and Girous, 
                    New York) 1998 for an introductory text. (2) Katherine Hayles, "Bodies out 
                    of Voices, Voices out of Bodies: Audiotape and the Production 
                    of Subjectivity," in Sound States: Innovative Poetics 
                    and Acoustical Technologies Ed. Adalaide Morris, University 
                    of South Carolina Press: Chapel Hill, 1997. (3) Steve Reich, liner notes for LP 
                    Music of Our Time: New Sounds in Electronic Music (Columbia 
                    Odyssey, New York) (4) For more on Georg Antheil's Ballet 
                    mécanique visit the site mainted by Paul Lehrman, 
                    www.antheil.org (5) Paul Lehrman, "Blast 
                    from the Past," WIRED, November 1998.  (6) For a basic review of how sound 
                    works in Flash, start with Working 
                    with Sound and Sound 
                    Objects: Controlling sound in Flash 5  (7) From the Rhizome 
                    Database statement for Orbital.
 
 
 TOP  |