Interactivity & Non-Traditional Productions

Selected Work

I research ways to create new design experiences for audiences. To do this, I work with dense pieces of software to create custom programs, audio systems, and interfaces. Often I use emergent technology in conjunction with old forgotten relics.

Computer Vision :

This is a video utilizing a depth sensing camera and some custom coding in Max/MSP to create 3D visuals of a space in real-time.

(2019) Feed:

I had the pleasure of designing this at the Cleveland Playhouse. The script is split into two halves; one half of which is in a typical script format and the other half is running a constant monologue of commercials that underscores the entire show. For this production, we decided that we wanted to do these monologues live. To aid in doing this, I created a custom Ableton based vocal effect setup that allowed for user-controlled effects to be added onto the voices of the actors. This process gave the performer’s more of a sense of ownership over their commercials, and, because of this, the commercials felt much more organic than if the underscoring was played through QLab and the A1 controlled the vocal effects at the board.

Audio Reactive Programming :

These are examples of audio reactive programming I created utilizing Jitter in Max/MSP. Using openGL one can create real time interactive audio visual experiences. In these two examples I wrote the code and the music to make small audio visual experiences.

(2017) Akron Haunted School & Lab:

This behemoth of a project happened in the Summer of 2017. Eighty-two individual point source speakers were hung across the 7 floors of 2 haunted houses. As you walk through, the content shifts from room to room sculpting unique environments and coloring each space.

(2018) Slow No Wake:

​In early 2018, I was a contributing artist in the event entitled Bricolage put on by the art company Maelstrom Collaborative Arts in Cleveland, Ohio. In it, as the contributing artists were paired together at random, I was paired with the analog projection artist Nathan Melarangro. We decided that we wanted to improvise in front of our audience live each night in a piece that we entitled Slow No Wake. To do this, I created an Ableton-based live composition tool that allowed me to create live sound design and music that matched his improvised visuals.

Custom Software:

I use creative show control strategies to achieve show specific effects. Often obscure software packages can create original design moments. I have controlled audio/visuals with cellphones, microphones, and body movements.

This is a preliminary ​Vocal effect software written in the Max/MSP environment for a show in the UCSD Wagner New Play Festival that is now canceled due to Covid-19. It makes use of the Spat~ suite created by IRCAM as well as an array of resonant filters.

This is a piece of software created to write generative music. It was used in collaboration with the music collective Suzo as well as in a public performance at UCSD. Each note is generated at random points in time and then patched into a series of feedback delays and filters. This creates a bed of sound that is reminiscent of stars. A microphone can be added as well for a similar wall of sound effect.