Thanks to Nancy Lowe from AS IS Center near Penland School, I am “new best friends” with Mark Boyd. Mark is a sound artist who records and amplifies the “voices” of plants, ants and flowing water: the realm of the tiny vibratory world. Talk about deep listening! Mark has been using electrode sensors on plants into a Volca Synth to listen to the electrical life force within the plant! He sent me recordings of his “biologues” with Bleeding Heart and Fern, and Dogwood. The playback presents us with a lot of fast-paced random sounds. Mark is interested in transducing this data into something people might listen to. Here is an excerpt from Mark’s Bleeding Heart and Fern biologues:
Ableton Live contains numerous tools for transposing/transducing/converting sonic data. An audio clip, such as the one you just heard, can be converted into midi clips; one that renders melody information, and another that renders harmonic information. So now we have audio information rendered as 2 packets of midi information. Within Ableton, midi files can be collapsed or stretched across a timeline and still maintain the integrity of the rhythmic intervals. Midi data can be assigned to a “voice” that feels representative of the sound artist’s impressions of the particular plant that is speaking. Midi data can be fed back into a synth such as the Volca to complete the circle.
The scaling of the time frames of the midi clips is exactly what is needed to help us “hear” the biofeedback from the plants. Doubling the length of the midi clip slows the overall “tempo” and helps us to listen into a kind of river of sound emitted by the plants. Slowing down allows us to tune into a rhythmic cohesiveness that is obscured by the frantic pace of the plant’s raw electrical impulses. We inject spaciousness into the mix in just the right amount, and it sounds like something is being communicated.
After finishing the new rendering of the data, I sent this to Mark:
He was ecstatic, over the top about all the possibilities of Ableton. He downloaded the Lite version and took off with it. We had some great email exchanges and he sent me samplings of his tests and experiments with the flora around his mountain home. Here is a beautiful example with a plant in Mark’s home:
I look forward to Mark putting together an orchestra of local flora in concert in the near future. In the meantime, I am enjoying dialoguing with another human being who is listening as deeply as I am.
TODAY, April 18th, iBoD is taking Durham to the National Water Dance! At 4 pm today, Jody Cassell will dance with the trees, the breeze, the sunlight and the water to iBoD’s Carnatic Water Music. The original plan was to create a watery container of sound, projections and flora through which Jody Cassell would lead the movement. Several other dancers including some of Jody’s young dance students were excited to perform this event at PS 137 in downtown Durham. Well, you know what happened! With so many cancellations, NWD leaders decided to go ahead with the dance as a streaming event. We will have Jody dancing to iBoD’s Carnatic Water Music at 4 pm this Saturday on Instagram Live from the safety of her home.
Here is some background on National Water Dance drom their website:
National Water Dance is a catalyst that encourages ongoing engagement between dance and the environment. Beginning in 2011 as a statewide project in Florida, it now boasts 65 locations across the United States—in 2018 including Puerto Rico for the first time. With each event National Water Dance recommits itself to the effort of increasing purposeful awareness to drive action on environmental issues, specifically climate change. Supporting the work of participants beyond our bi-annual event, National Water Dance spreads the word on the environmental issues they are tackling through monthly newsletters and social media. Our goal is to realize the power of dance as an engine of social change.
In the spirit of this mission, Jody suggests, through her movement choices, elements of this “engine” that are not as visible as issues. Stillness, slowness, wonder, and (dare we say it!) JOY in being right here and right now and moving! The soundscape Carnatic Water Music articulates and ensounds a few of the ways that water can move! To hear the entire piece go to Bandcamp/dejacusse.
As this large group of dancers in 65 locations across the country create a wave of healing energy directed toward Earth and Water, your attention will amplify the energy beyond all possibilities.
Join us for a five minute healing prayer on Instagram (@movinggrace) or through the National Water Dance Website. If you cannot join us at 4, please bear witness to the dance on Instagram TV. Now, more than ever, it is important to mind what we ourselves are giving our precious time and attention to!
My curiosity about sound is completely engaged by exploring modular synthesis. So far my understanding is often inarticulate and mystified! But thanks to Suzanne Ciani, True Cuckoo, Andrew Huang, Ultrabillions, Hark Madley, Lisa Belladonna, Caterina Barbieri, Moogfest, Bram Bos, and Kim Bjorn’s book Patch and Tweak, I am evolving a different way of creating soundscapes and perceiving the world. This is the stuff of life! Waveforms modulating waveforms, waveforms shaping waveforms, waveforms reflecting, refracting and bouncing around and through us. Energetic matter begins and ends on a wave.
I am focusing my Artists Residency here at home on improving my mixing skills and building a sounding board. The mixing skills are put to the test making the recording of Carnatic Water Music that iBoD will release in the next week. As I mixed this recording I received helpful suggestions from tutorials by Jason Moss, HarkMadley, Mathew Weiss. These skills are a forever work in progress. As for the sounding board, there are currently three main ingredients: Elektron:Model Samples as main sequencer providing beats/patterns and midi triggers to the Behringer Neutron. Audio out from both of these units into Audio Tracks in Ableton Live. Ableton will provide drones, loops, and AAC/EG clips which can process audio from either unit. I can do Master recordings in Ableton as well.
Even though I want a modular system, I will work with what I have now, and learn, and be ready when my modular system appears. (Make Noise modules are the ones that I want- doo doo do do)
The Model Samples and I are getting on fairly well. I am learning the architecture of the menus, watching people perform with it to see what key combos they use, and setting up some patterns. The samples available “in the box” are very cool and I am curating my own samples as well. Every sound is potential material so it is daunting.
The past few days, I experimented with some patch ideas in the Behringer Neutron. I have gotten alot of growling out of the synth, but no sound that I liked. There is one simple patch I use: the Sample and Hold into Delay Time. When the Delay Mix knob is raised and the S&H knob is turned up, there are lots of odd, random pitch artifacts that I enjoy hearing. Today I patched the Osc Mix into a Mult, then ran Mult 1 to the OD(overdrive)IN, and Mult 2 to Pulse Width 2. Tuned the oscillators to consonant pitches. Slowly turning the Osc Mix Knob opens a whole realm of timbres. When the OM knob was all the way to one side the tone could be made clear and bell-like. With the Oscillator shapes in the square or tone mod shape, the Pulse Width knob seems to act as a filter.The Mod Depth and Envelope Depth can be brought in. This is where I am not sure what is happening – there are changes in the timbre of the tone from the synth. And what exactly is depth? There is alot to play with depending on where the Osc Mix dial is tuned in.
The third part of this is creating Audio Animation Clips/Envelope Generators within Ableton. Envelopes shape the amplitude and modulate the pitch of the sound. Audio Animation allows the Envelope parameters to move over time. Here is the post on how audio animation can be created in Ableton: https://wp.me/p5yJTY-vL I use filters to sculpt out harmonics and add texture to the sound of the Model Samples or the Neutron. So far, I am experimenting with banks of filters to sculpt out or boost particular harmonics then perform a finer tuning with some EQ. I am listening for a diverse sonic spread, then tuning it in, then spreading, and finally fine tuning.
Right in the midst of the most recent Mercury Retrograde, I decided to dive into MAX MSP, a visual computer coding program for controlling sound and light for performance. After downloading the software, I started a class online and was working with some patches when my computer audio stopped functioning. No sound out of the computer. Then the computer and sound card stopped talking. All of this right before an iBoD rehearsal when we were recording Carnatic Water Music.
Using the Windows Troubleshooter, I discovered the problem “audio services not responding” and that this problem was “not fixed”. Online, there are multiple fixes for this message. After cancelling our recording session, I tried all the suggested fixes several times – from inspecting the Services to make sure Windows Audio and Windows Audio Endpoint and all their dependencies were automatically running to entering very specific commands into Command Prompt as Administrator. The first thing I did was update the ASIO4ALL audio driver, so no problems there!
After several days of trying different fixes, I was able to get the computer and sound card talking again! Ableton Sets and Projects were now audible! Yayyyyy! But the computer would not play audio WAV files. Outside of Ableton, audio services still not responding. Finally, I uninstalled the ASIO driver and uploaded the driver for the soundcard. I have a Native Instruments Komplete 6 soundcard, which has been a great device. (I had audio dropout problems with the NI driver about a year after I purchased it, which was when I switched to the ASIO driver and all was well.) Well, changing back to the NI driver solved the audio problems completely and I am back to sounding again!
A friend mentioned Mercury Retrograde as I was working through this process. Dang, I forgot about that current astronomical phenomenon. If I had remembered, would I have done anything different? As things turned out, it is very good that I did not! While I got thrown off of MAX (for the moment) I redirected my energies toward creating synth sequences in Ableton. Since purchasing the Behringer Neutron, I have been unsuccessful in getting Ableton set up as a sequencer for the Neutron. The Neutron has processed audio signal, but never midi signal. Low and behold the NI Komplete 6 driver allowed Ableton to see the midi ports for the Neutron. Suddenly, I was hearing the synth voice and all the modulators. When I made a patch or tweaked a knob, the sound was changed as I expected it to be! WoW! I feel like this is the first time I have heard the instrument’s true voice!
Today I am working on a soundscape for the next Human Origami Jam at ADF Studios in Durham on December 6. Very excited to finally get going with the Neutron.
Lisa bought a new guitar! A John Suhr limited edition commissioned electric guitar signed by the maker in a faux alligator hardshell case. The top of the guitar is quilted maple and looks like rippled water. Lisa bought the guitar because it’s voice eclipsed the sound she was carrying around in her mind. She said she had this jazzy sound in mind with rhythm (swingy, danceable) and a clean, clear tone when plucked (like George Benson). The Suhr guitar has a lovely tone with crisp, clean edges and bell-like shape. The sound the Suhr guitar planted in Lisa’s earbrain is more “New Agey”
A few weeks back, I sent Lisa a thumbdrive with recordings of our sessions since June. She reports that the recordings were not helpful to her as she couldn’t pick out her voice from the whole soundscape. This is good to know- the recordings give me a lot of information, but not so for Lisa. I know she listens to music by turning it up very loud in her home, so I asked if she did the same with the session recordings. She explained that she has sound reference files in her brain that pick up on familiar patterns associated with the song she is listening to. Without these references, Lisa is less able to make sonic sense of what she is hearing.
Our September 28 2019 session focused on the new guitar and what it brings to our pallette. And we played in a different relationship today. Instead of Lisa’s guitar through the Neutron, we played on separate channels. Lisa wanted to hear her new guitar clearly since she is just learning it, so I played the Ripplemaker through the Neutron. In this configuration, Lisa leads the way, while I bring interesting underpinnings into the mix.
Listening back to the recording, I think this is another way for us to play together. Our collaboration becomes more like intermingled solos, so the impact of our playing together is indirect rather than direct. Our voices are tandem rather than merged, and we can respond to each other. One question is how to create useful audio reference patterns for Lisa? She said that she couldn’t hear the recordings in the thumb drive because they were too removed from what we are doing currently. So it seems possible that if she listens to a recording from the most recent session, she could create new reference files. We will try this out.
The October 5 session is when things came together. Lisa brought another guitar – a 17″ wide arch-top Kay guitar which she describes as the kind of guitar you would find in the Sears catalogue in the 1950s. She played that and the Suhr while I created morphing streams of sound sequenced by Ripplemaker and modulated by Neutron through Abejusynth Station modules. The quality of the sounds of the sequence can be altered within the Ripplemaker, then in the Neutron. Then the audio signal from the Neutron goes through an Ableton audio track, which can then be sent through and altered by the Abejusynth Station AAC/EG modules. (For more info, go here: https://wp.me/p5yJTY-vL). Any of these Ableton tracks can go through delay send and a reverb send. So there is a whole lotta modulating going on!!
After our October 12 session, I am very excited about our playing as intermingled soloists at 919 Noise Showcase on October 30. We ran ourselves through my Roland Eurorack mixer (Thanks, Jim!) so I could balance the sound. Then I recorded into 2 H6n tracks and in the room. We decided to start with a wave of sound and then whittle it down. I was not sure this was working, but listening to the recording, I decided we need to just listen close and have faith that it IS working.
Here is a mix of the 2 H6n tracks AND the room recording. This seems like an interesting way to capture sound recordings in the SunRa Room. That said, this mix has too much synth and not enough guitar, and we will fix that so the blend is better in the future.
My quest to synthesonize Ableton Live has taken an exciting new turn. Last Sunday, we discovered that by micing The Bells at the Central Park School Soundgarden, I can run that sound through Ableton and into the various synth modules and FX racks I am building. What happens is that the Abeju Synth Modules and FX Racks capture most of the harmonics that arise from Eleanor’s bell playing. The harmonics can be shaped by envelopes and attenuation and, of course, granular synthesis. My goal is to gradually shape the bell harmonics into a watery stream sound. This will be part of the soundscape for The Place ReSounds of Water (TPRSW) on April 14th at 4pm for SITES Season 2018-19.
When iBoD first started playing with The Bells, I recorded and analyzed their harmonic content. These bells are former compressed air tanks with the bottoms cut off, so the metal is not pure, it is some kind of alloy. This translates to lots of harmonic AND enharmonic content! A pure metal would render more pure harmonics. These pure harmonics are pretty, often beautiful, but my ear grows tired of the stasis of it all. The idea of purity in all of its forms is an illusion that leads to much misunderstanding and anguish in the world. Think about what striving for purity has given us: genocides, fascism, chronic autoimmune diseases, disconnection from and attempts to conquer nature, diminished empathy, and on and on. It is my prayer that riding and faithfully playing All the en/harmonic waveforms will encourage evolutionary growth. That is what I am going for!
TPRSW is my first attempt to sync up with the National Water Dance. My timing is off as this is not the year for National Water Dance, however I am hoping this will kickoff some interest for 2020. The idea for TPRSW is to give prolonged loving attention to water in the form of sound, light and the liquid itself. The soundscape will consist of Eleanor Mills playing The Bells, dejacusse aka Jude Casseday capturing and playing the en/harmonic waves from The Bells and morphing them into a watery feeling soundbed. Then Susanne Romey will play Native American flute over that for a while, then we start the wave again. The movers will pour water from vessel to vessel. An altar of flowers may be built. The whole thing is a mystery.
Our location at the Soundgarden at Central Park School gets full afternoon sun, so the visuals might include sparkles and shimmers of water. We could be lit up! If it is overcast, the air will be moist and the sounds of water will carry more clearly. If it threatens rain on Sunday, we will do it on Saturday instead! Or, perhaps, we will figure something else out and perform as it rains.
When I was a child, we often visited our grandparents in Elkins WV. Elkins is home to the Mountain State Forest Festival, and is my birthplace. My Mother’s family has a long history with Elkins. Her grandfather was one of the first mayors and one of two doctors after the town’s 1890 incorporation. I am not sure how my Dad’s mother got there. Mamaw lived in a brick row apartment with a porch and stoop to play on. And she lived one block from the volunteer fire department.
When I slept over with Mamaw, there was always a fire in Elkins, sometimes two. The volunteers had to be called in from all over town, and what called them was the longest, most mournful sound my young ears had ever heard. As loud as it was (remember we were one small block away) the siren also sounded ghostly. It went on and on and on for an eternity and then it stopped! A lovely silence would fall and gently wash away the residue of the wailing. If it happened at night, I would return to sleep; by day, it was back to play. Either way, the siren always elicited a jolt of free-floating anxiety.
The Mountain State Forest Festival takes place the first weekend in October in Elkins and has for 85 years (with a short hiatus during WW II). This Festival was a highlight each and every year of my growing up. We got out of school for two days, traveled through the gorgeous colors and crisp fall air to spend several days with carnivals, exhibits, parades and pageantry. One of the parades took place on Friday night and involved 100 firetrucks sounding their sirens at the same time. The Fireman’s Parade attracted fire departments from all over West Virginia, and into Virginia and Maryland. The trucks would line up at one end of town and slowly make their way down the main street blaring the siren song of their station, their truck. The sound of 100 firetrucks calling their warning song together cannot be described. People flocked the sidewalk, laughing, trying to talk to each other over the din. My brother Matt is famous in our family for having slept through the Fireman’s Parade when he was a babe. Even back then, I enjoyed the interplay of the various intervals that make up a siren song.
A few years ago, my cohorts from iBoD (idiosyncratic Beats of Dejacusse) were discussing ideas for soundscapes. The one sound artifact that really stands out in the urban growth we are experiencing in Durham NC is the frequency of emergency sirens. This became the basis for an iBoD piece called The Sound…of Sirens. One online resource said the intervals of sirens telegraphed who’s coming: the police are a perfect fifth, ambulance is a fourth, and fire trucks are a whole tone. I designed the soundscape with those intervals. We all started with the basic intervals, and as the piece went on, we threw different intervals into the mix. The ending is a big crescendo and all out except the tail of the reverbed voices of the scape, which I turn up to a final fading shriek. We played the piece at a few venues. I thought of it as a novelty song.
I talked about all of this in an interview with Margaret Harmer, who produces electronic music as Shifting Waves. Margaret is producing an album of work from 15 to 20 women electronic artists from all over the world. She asked each of us to think back to a sound in our childhood, to find the story around that sound, and bring it forward into a piece. (I actually added that last part, Margaret did not say the story had to be about the piece for the album, and it sure did flow that way for me.) Here is a link to the interview.
I took the soundscape for The Sound…of Sirens and began to analyze it harmonically and timbrally. The piece was sculpted from thick resonant voices (several synth pads and strings). This allowed me to carve out the movement of the sirens, the doppler effect of approach and recede, the abruptness of a nearby siren suddenly starting or stopping – the psychoacoustic impact we experience in our communities. Now called Song of Sirens, the piece was a fountain of siren voices overflowing and receding. There are several short repeated interludes during the first section. Several crescendos and several interesting places where the sound drops out leaving space in the front of the mix. This is most obvious when listening through headphones. This has peaked my interest in how we define the sonic space a piece takes up, and how to keep the full space alive when the sound recedes.
Siren’s song in mythology is characterized as an intentional “luring” of sailors onto the rocks. This sounds like one side of the story to me. Who was hearing and for what end? Was the siren song seductive, plaintive, demanding? Was it the call of grey seals, baying and mournful, resounding in the range of the female voice, a voice the sailors had not heard in years? Perhaps the sailors drove themselves into the rocks looking for women to rape. There are many possible scenarios when all points of view are considered.
I wanted to put an intention of comfort and nuturing from female voices into Song of Sirens. How interesting that modern day emergency sirens call out warning, answer your cry for help, or pursue you – all at once. How to embody all of this while flipping the mythology of blame the women. So I recorded Trudie, her daughter, Sheila, and three granddaughters singing phrases of Brahm’s Lullaby and wove them in and around the siren soundscape.
We are creating a new mythology as our brains and conciousnesses go through an extraordinary evolutionary shift. The reptillian brain – the one that fights or flees – is softening into the polyvagal brain. We are moving from survival of the fittest to survival of the kindest. Feminine consciousness knows how to be kind, not just benevolent. As the Song of Sirens raises the death knell of the reptillian brain, grandmothers, mothers and granddaughters sing a soothing lullaby swaddling the panicy cries.
Song of Sirens will be released as a track on Voices from Eris, produced by Shifting Waves studios. Stay tuned for more on fundraising and release date. I appreciate your listening!
Moogfest 2016, which took place May 19 – 22 in Durham, was a mind-blowing and inspirational experience for me. Last Fall, while selling my old instructional drumming CDs to the now-defunct Nice Price Books, I was talking to the owner about my new love: electronic music. He said, “You must be super excited about Moogfest coming here!” “Oh, yeah”, I responded, knowing I should be excited but just not feeling it yet. A few years earlier I wanted to go to the festival in Asheville, NC when Brian Eno was featured. But then I read how you spend all this money on a ticket and might not be able to get in to see what you came to see. So I knew about how the tickets worked, and that it was a celebration of Bob Moog, a synthesizer pioneer. The Moog Factory is still a fixture in Asheville, but Moogfest was coming right to my front door.
I was still feeling ambivalent in April and Moogfest was 6 weeks away. One thing I had decided – I wanted to be involved musically – so I started planning a Post-Moogfest event for the final day after everything “official” was over. (See post: http://wp.me/p5yJTY-ci) Then a volunteer application came my way, I filled it out and attended my first volunteer meeting. I met Wilson, Hugh, Robin, Ilsa and several other sweet, friendly folks who were psyched for the event. Bianca Banks, the volunteer coordinator, gave us postcards and Moogfest stickers (everybody LOVES stickers) and a welcomed us to the Moogfest family. Sweet!
The only acts I knew in the line-up were Laurie Anderson and Sun Ra Arkestra. By this time, Sun Ra Arkestra had cancelled, so I started YouTubing the artists to get a taste of what they had to offer. I started with the women artists: Julianna Barwick, Grimes, Suzanne Ciani, Grouper, Julia Holter, Laurel Halo, Olivia Block, Paula Temple. I did not get very far in this exploration before Moogfest was upon me and I just had wing it.
The first day, I worked guest check-in with Michael Jones (or Jones Michael, his producer moniker: check out his Soundcloud – https://soundcloud.com/jonesmichael), Nico and several other young musicians who told me about groups they were excited to hear. Volunteering took 18.5 hours of the weekend, and got me free admission into the festival – way worth it. I learned that hospitality is not my skill set (My partner, Trudie said, “I could have told you that.”) I learned that there are lots of folks, young and old, poor and rich, out there creating vibrations in the form of music and sound. I learned that people who come to Moogfest are – for the most part – friendly, open and excited about the prospects of technology and music making.
Luckily, Jim Kellough recommended several performances to me on the first night that were fantastic. His first recommendation was Silver Apples, a staple of the NYC scene since the sixties. Silver Apples was an early electronic duo who played the soundtrack for the moonlanding as it was broadcast on a big screen in Central Park in 1969. Now Silver Apples is just Simeon (his drummer died in 2005) and he really rocks the synthesizers. Here is a picture of Simeon with The Soundman AKA Christopher Thurston at Motorco the night of his performance:
Christopher and Silver Apples, Motorco, May 19, 2016
After this show, I headed over to see the best music of the whole weekend. Arthur Russell’s Instrumentals was inspired by the nature photography of Yuko Nonomora, and was only performed five times in Russell’s short life. The group, playing under the direction of Peter Gordon, was comprised of Russell’s collaborators and cohorts, including Peter Zummo, Rhys Chatham and Ernie Brooks. The piece was jazzy, funky and took the listeners on a fabulous journey. My favorite part was Peter Zummo dancing around the stage and gently clapping his hands whenever the trombone had a musical hiatus. Their performance left me curious to check out more of Russell’s work.
Moogfest is all about synthesized sound. So on Saturday, I headed down to The Carrack to hear Antenes, who crafts old phone operator switchboards into sequencers and synthesizers. She performed on her DIY synths for a half an hour and then did a presentation on how she came to create these particular instruments. I loved the deep sweeps and blips and bloops she carved out of various oscillating waveforms. Next stop was the Pop-Up Moog Factory, where employees were building actual Moog Synthesizers right before our eyes. The employees worked at four stations performing assemblies and passing them on to the next table. By midday Saturday, they had assembled 14 Minimoog Model Ds. The factory was full of a variety of synths hooked up to headphones so people could play and experiment to the ear brain’s delight. I had a fantastic several hours there, and left feeling like I really need a synth to add to my setup.
Then I checked out Critter and Guitari, who were in a geodesic dome tent outside the DPAC. These Booklyn-based musician entreprenuers have created adorable little synthesizers that are just my style. I enjoyed playing with the Moogs, but they are expensive and heavy. (Dang, I do not need anymore weight in my setup with a 12″ QSC K Speaker to haul around.) I enjoyed jamming with the guys , the other peeps, and the train that passed by. Their Organelle allows you to dial up a variety of sounds, play them polyphonically on a little wooden button keyboard, and tweak the sounds as you go. Neat! In my fantasy, they offer to give me one to play as a sponsor of ibod when we go on our sound sculpture tour. Wouldn’t it be nice…
I was anxious to get a good seat for Laurie Anderson’s Saturday afternoon performance, so got there waaaay early only to discover a long line snaking around The Carolina Theatre. I got in it only to discover the line was for a talk by Jaron Lanier, whose name I did not know. The guy in front of me did not know him either, but he figured “He is the keynote speaker, he must be good!” As it turned out- he was right! Jaron is a musician, virtual reality geek, author and incredible human being. He started his talk by playing the khene, a Laotian mouth organ, that he said is a “digital” instrument thousands of years old that could have inspired the invention of computers. Here is a YouTube video, where he plays this instrument in his own amazing way:
His message was wonderful and optimistic. He said we need to “will away” our obsesssion with war, combat and all things military. He advocates a movement toward kindness and beauty as guiding values in technological development. He asked VR game makers to use the technology to engender empathy. What I heard was – let us play games that engage our emerging polyvagal brain rather than continuuing to stir up our shriveling reptillian brain. Jaron Lanier is one gorgeous genius, and I was uplifted and inspired listening to him.
Next up was Laurie Anderson, who grabbed her electric violin, slung it over her shoulder and and filled Fletcher Hall with deep sweeping harmonics that made my heart pound. She moved toward the audience as she continued playing, looking right at us. This connecting more openly with the audience is a shift in her performance aesthetic from times I have seen her over the past twenty years. The next day, she talked about “seeing the audience” during her presentation/interview. While I enjoyed her performance, I was mesmerized by the retrospective talk about her work on Sunday. I love hearing and reading about artistic process. It is extremely intimate discourse, which is why many creatives are reluctant to share it. Laurie gave us a glimpse into her process over the years, and for that I will be forever grateful.
She spent a good bit of time talking about a recent work Habeas Corpus and how the piece evolved into an illumination of and a step toward healing the horrors and injustices of Guantanamo Bay. The work was presented in 2015 in NYC and is based on the experience of Mohammed el Gharani, the youngest detainee at Guantanamo Bay. He was sold to the US at the age of fourteen, kept in solitary, subjected to torture, and finally released by a US District Court judge for lack of evidence. He was held for seven years. The performance installation included a plaster cast chair the size of the Lincoln Memorial. Mohammed’s full body image was projected via a live video feed from Chad, where he now resides. He sat in the chair and told his story. The audio was one way only to protect Mohammed from hearing any personal attacks from the American public – there was concern that those Americans still blinded by their own fear and ignorance might attend the installation to berate him. He had suffered enough at American hands already. The video feed was two way, so Mohammed could see the audience. The most moving thing Laurie shared with us was that many of the attendees came forward and mouthed “I am sorry” to Mohammed’s projected image. For more on Mohammed el Gharani and Habeas Corpus see this link:
Laurie Anderson echoed Jaron Lanier’s thought on the necessity for kindness, empathy and beauty as hallmarks of our creative relationship with technology. Both pointed toward the potential for technology to help us connect, see, listen to and understand each other even if we do not agree.
Laurie and Lou Reed, her husband who died of cancer in 2013, came up with three rules to live by which she shared with us: 1. Do not be afraid of anyone. 2. Have a good bullshit detector, and learn how to use it. 3. Be tender with life. Afterwards, I could only remember 1 and 2. That is because I have issues with tenderness. Tender feelings make me feel vulnerable. Gotta work on that.
There is lots more to write about, so many encounters and experiences packed into 4 days, 40 venues and nearly 300 speakers/performers/presenters. Moogfest was so much more than I ever expected – my world expanded several times over. And the best way to top it all off was to play with my cohorts before an exclusive and appreciative audience. Here is an excerpt from Adrift in a Sea of Bells, one of the pieces we performed in the soundgarden following Moogfest:
This experiment began with a rather dubious YouTube video about the “11th harmonic” and its power in breaking up cancer cells. The video is about the Rife Machine, which was an invention from the 1930s purporting to cure many diseases. Royal Rife was the scientist and inventor who “discovered” frequencies that could interfer with the frequencies of diseased cells. The narrator of the YouTube video, stated that the 11th harmonic was the frequency that disrupted cancer cells. About a week after I started this post, I found a TED Talk along this same line:
What we are learning from quantum physics about how the Universe is put together lends quite a bit of credence to the idea that frequencies can disrupt disease. Oscillating frequencies make up the entire spectrum of “all that is.” When these frequencies interact with consciousness – “being” happens. Our singular awarenesses collapse the waveforms into the many points of existence – the mix of all our singularities creates what we call “reality”. The famous physicist Erwin Schrodinger put this idea in another way when he said, “The total number of minds in the Universe is one: In fact, consciousness is a singularity phasing within all beings.” Oscillating frequencies engage with each other through constructive (in phase) and destructive (out of phase) interference (or, as I like to call them – engagement) patterns. Thus the fabric of reality is an oscillating organism of frequencies engaging, changing and disengaging with each other. Our brains stabilize the whole thing so that we can navigate and participate in our lived experience.
Both of these videos assert that a harmonic relationship created by a low tone and a higher tone is necessary to disrupt diseased cells. In both cases, the necessary frequencies equate to an extreme number of oscillations. Dr. Holland said that frequencies needed to be around 300,000 to 400,000 hertz in order to destroy cancer cells. While these frequencies are waaaay outside of the audio spectrum, there is an organizing principle that allows for the possibility that lower audio frequencies might influence healing. And that organizing principal is – the octave. Whatever frequency you start with will always return “home” when it doubles. It is itself again. For example, middle C on a piano is about 262 hz, double that to 524 hz and you are at C again. This creates a resonating fractal that repeats on and on into infinity.
The harmonic overtone series, which is the basis for most everything we hear musically, is built around this doubling principal. As we add more iterations of the fundamental frequency, we create more overtone relationships. Using the middle C example again, adding 262 hz to 524 hz gives us 786 hz, which is G or a fifth above C. Add 262 hz to 786 hz and we get 1048 hz which returns us to C again. Now we are two octaves above our fundamental frequency Middle C, AND we are at the 3rd harmonic. By adding 262 hz eight more times we reach the 11th harmonic, which is 3114 hz – G in the fourth octave above middle C. (For more on harmonic overtones and their impact on our cosmic existence check out Hans Cousto’s book The Cosmic Octave.) Now I can create an audible 11th harmonic by combining a fundamental frequency and the fifth degree of that frequency in the fourth octave above that frequency. So I decided to make a leap of faith into the realm of the cosmic octave, and create a soundscape that hinges on an 11th harmonic and the healing secrets that it may hold.
Folding/Unfolding: The 11th Harmonic is built on a tetrachord of fundamental tones – CEGB accompanied by their 11th harmonic companions – GBDF#. The tones are 4 octaves apart, so this is not an interval you are accustomed to hearing. I chose 6 instruments and created patterns with these unusual intervals. As I thought about how to voice this harmonic, I identified three choices :1. alternate between the fundamental and harmonic in a variety of rhythmic patterns all on one voice, 2. have one voice sounding just the fundamental and a different voice sounding the harmonic, 3. since the 11th harmonic is a fifth in the fourth octave and the two octaves below the fourth octave also contain fifths (according to the overtone series), then I could vary the patterns with some fifth reinforcements in those lower octave. The second choice was very monotonous and weakened the presence of the 11th harmonic, so I went with the other two as my basic structure.
This soundscape will be performed tomorrow, May 15th from 2 to 4 pm as accompaniment for Glenna Batson’s latest Human Origami workshop. This workshop is subtitled Partnering with Paper, Exploring the Muse. Joy of Movement Studio in Chatham Mills is hosting the event. In addition, to the featured 11th harmonic, I will use the audio folding techniques I discovered during the previous Human Origami workshop.(See blog post – http://wp.me/p5yJTY-c9)