New Book: Sounding Human

My new book, Sounding Human: Music and Machines, 1740/2020, is out now from University of Chicago Press. Much of the music discussed in the book is available on the accompanying Spotify playlist. Read on for a (tweet-per-track) tour through the playlist and some audio/video examples not available on Spotify!

Introduction

Track 1 Mozart, Fantasie in F Minor K. 608 (1791) illustrates a contrast between modern and 18th-century intuitions about music & machines. It was composed for a mechanical organ clock, and as musicologist Annette Richards pointed out, 20th-century critics dismissed the machine as cheapening of Mozart’s music. In its initial reception, the music & its automated performance went together for a heightened awe-inspiring display (one witness described “a thousand varying emotions”). Listen from 8:05 for the increasingly dense and dramatic fugue

Track 2 Kraftwerk, “The Robots” (1978) features the use of a vocoder to turn the voice – and human speaker – robotic. Florian Schneider described their music “expos[ing] the mechanical and robotic attitude of our civilization”

Track 3 Afrika Bambaataa & The Soulsonic Force, “Looking for the Perfect Beat” (1983) illustrates, in words of Tricia Rose, an “already having been robots” for Black people under slavery/capitalism and resistive taking control of the robot guise. Listen at 5:50 “we are the future, you are the past”

Wendy Carlos, synthesized finale from Beethoven’s Ninth Symphony (1971). The vocoded “Ode to Joy” is commonly heard as voices drained of humanity, but for Carlos and collaborator Rachel Elkind it was the sound of synthesizer gaining it by singing. Listen, for example, at 5:55

Track 4 Nona Hendryx, “Transformation” (1983) another alternative to vocoder = robot, Hendryx’s use can be heard to sound the voice of mother nature, dissolving natural versus tech/artificial. Listen ex. at 1:58 in the context of the song

Track 5 From Giovanni Battista Pergolesi, La serva padrona (1730s), the main characters mimic a clock bell and bass drum as the sounds of their hearts/love, to which they direct each other to listen, ex. at :51-1:10

Tracks 6 & 7 Queen of the Night aria from Mozart, Magic Flute (1791) & android Olympia’s aria from Offenbach, Tales of Hoffmann (1881) illustrate leaping vocal virtuosity that became sounding inhuman/like a machine. Listen @ :45 on Track 6 & 1:25 on Track 7

Trk 8 From Mozart’s Magic Flute (1791) at :38 – in contrast to the clockbell of a loving heart in La Serva Padrona (trk 5), bells here make slaves dance automatically, conflating the mechanical and Blackness in a “racist technopoetics” (Chude Sokei’s term)

Chapter 1 Becoming Android: Reinterpreting the Automaton Flute Player

Track 9 From opera Pirame & Thisbé (1726), the melody at :43-1:40 was sung in the play L’Oracle (1740) by a character pretending to be an automaton that can only “parrot…at random,” raising questions about being moved by a machine & being human through its singing

Chapter 2 Hybrids: Voice & Resonance

Tracks 10 & 11 Last Magic Flute example: classic interpretations of these tracks illustrate modern hearing of human versus mechanical sound in voice versus bells

Track 12 Alessandro Striggio “Chi fara fede al Cielo,” appears as hydraulic organ music in a 1615 treatise on machines by Salomon de Caus, and exemplifies what might have sounded as Orpheus’s mythic music in the mechanical grotto featuring Orpheus playing a viola da gamba

Track 13 This Prelude for viola da gamba by Marin Marais features letting strings ring, a technique criticized for “stand[ing] in the way” of melody. Compare beginning (continuous melody) to ~1:35-2:00 (technique used to let notes ring while playing others)

Track 14 This viola da gamba piece by Marin Marais imitates the style of a plucked lute, and captures a perception of lute sound quickly decaying – “tic tac” – rather than ringing out

Track 15 In contrast to the lute-imitative piece, this viola da gamba piece by Marin Marais imitates bells, makes extreme use of letting strings ring (ex. :40), and suggests what a defender of the viol against the violin’s rising popularity meant by its ability to reconcile voice & resonance

Track 16 From a violin sonata by Michele Mascitti (1706), illustrates continuous “voice” sound of violin but played on viola da gamba would illustrate a “sparkling sound” combining melodic voice w/ harmony of resonance (according to a viol defender in 1740)

Track 17 This François Couperin piece for harpsichord features a technique of delaying a melody note to create the effect of a crescendo (an illusion of starting on time and growing louder). This was a means to “give soul to the instrument” and makes the harpsichord exemplary not of lifeless mechanism but of music entangling matter and mind. Effect occurs at :56, 1:02 and most notably 1:17.

Chapter 3 Analogies: Diderot’s Harpsichord & Oram’s Machine

Trk 18 “Four Aspects” by Daphne Oram (1960). Oram used to illustrate crescendo requiring more than just increased volume: “To get a real feeling of crescendo one has got to…put more & more high frequencies into the tone color as well.” Listen 2:16-3:37

Trk 19 “Contrasts Essconic” by Daphne Oram, performed @ First London Concert of Electronic Music by British Composers (1968). Its contrasts were in mood & sound sources, which included piano, tone generators, musique concrète & the debut of “Oramics” the “graphic sound” system Oram invented for hand-drawn control of electronic sound. Oramics most likely heard at 3:56-4:50 and 5:40-6:43.

The 1968 concert also featured Peter Zinovieff’s “Partita for Unattended Computer,” which got media attention as the “first concert performance ever by a computer.” Featured in this Tomorrow’s World episode – what became sound of “computer thinking” @ 3:40 https://www.bbc.co.uk/archive/peter-zinovieff–early-electronic-music/zmyf7nb

Excerpt from BBC radio program World at One (1972) in which Daphne Oram demonstrates her “method of translating visual wave patterns into musical sound, in a way that retains the human element” https://research.gold.ac.uk/id/eprint/26356/27/011_Oram_1972_Interview_Excerpt.mp3

Chapter 4 Personifications: Piano Death & Life

New York Times 2012 coverage of pianos going to the dump & responses to it start Chapter 4’s inquiry into personifying pianos:

Annea Lockwood, Piano Burning (1968) https://www.utne.com/arts/bonfire-of-the-ivories-visualize-your-piano-burning/

Chapter 5 Genres of Being Posthuman: Chopped & Pitched

Track 20 Calvin Harris and Rihanna, “This is What You Came For” (2016). The “yooo ooo ooo”s were the 2016 sound of “people don’t want to sound like people anymore” (according to one music critic)

Track 21 Kiiara, “Gold” (2016). Producer Felix Snow described “manipulating vocals to sound like a computer and vice versa” as “the first tiny step toward the computer doing all of the singing”; Kiiara observed, “it’s a sample. That’s all it is.”

Track 22 Zara Larsson and MNEK, “Never Forget You” (2016). In contrast to pitched and chopped lead vocals, processed vocals in instrumental textures ex. 1:07-1:33

Track 23 T-Pain featuring Ludacris, “Chopped N Skrewed” (2008). This 2001: A Space Odyssey parody really sums up autotune anxieties ca. 2008

Track 24 Cher, “Believe” (1998), the breakthrough hit for the sound of autotuned vocals frequently characterized as “robotic”

Track 25 Flume featuring kai, “Never Be Like You” (2016) exemplifies ca. 2016 EDM-pop’s taste for “feathery” female vocals, ex. :19 and 3:30

Track 26 MNEK, “At Night (I Think About You)” (2016), in which producer-singer MNEK manipulates his own voice with and without the computer, as he discusses in this interview esp. at 3:05 https://www.bbc.com/news/entertainment-arts-36744245

Coda: Learning Machines

Track 27 David Cope, “From Darkness, Light” (2010) illustrates one pole of AI in the socio-musical imaginary, with anthropomorphic name Emily Howell emphasizing humanlikeness and implying its generalization to autonomy, etc.

George Lewis and Vijay Iyer in Concert (2012) with Voyager, which they discuss @ 36:30, foregrounding both its need for help to function and capacity for disobedience (a sociomusical imaginary for AI between the extremes of autonomous or “just a tool”)

Track 28 Laetitia Sonami, “Breathing in Birds and Others” (2017) stands in for this artist by Sonami in 2013, where she demonstrated her bird with magnetic coils (47:40); this talk was my introduction to neural nets https://archive.org/details/sonami

Trks 29-30 are from Dadabots, Can’t Play Instruments, Vol 1 (2021), an album of neural network-generated hardcore punk that continued the duo’s front of humanless musical creation (while elsewhere databots discuss “using the neural nets as an instrument”)

Track 31 Holly Herndon, Jlin, Spawn, “Godmother,” PROTO (2019) lent itself to idea of AI acquiring sentience

Tracks 32-33 Jlin’s “Expand” (2015) & “Embryo” (2021). Listening to these tracks clarifies “Godfmother” as the sound of a model trained on Holly Herndon’s voice & style transfer to Jlin’s music

Jennifer Walshe and Memo Akten, Ultrachunk (2018). Walshe describes performing the neural network trained on her voice as working with an “alien intelligence”

Hexorcismos, Transfiguración (2020). Another example of alternative to likening/comparing AI to human, Hexorcismos describes “approaching the Neural Network as an alchemical technology,” which aptly characterizes the process of style transfer

Tracks 34-37 Arca, “Riquiquí” (2020) & from Riquiquí; Bronze-Instances(1-100) (2020), the album of its 100 remixes via AI – or in her term that steps outside the usual frame of human comparison to contemplate machine capabilities, “intelligent sentience”



Leave a comment