“ontextC” – Technical Diary 2

What happened so far?

Aside from a crude mockup in Max MSP, a diagram helps envision the signal flow and processing points of the plug-in now. The diagram is also quite a handy tool to identify challenges, as it lays the main idea out in a layout that is simplified, but representative of the core idea. Parameters have been defined and narrowed down further.

I have also been provided with copies of all three volumes of Electronic Music and Sound Design – Theory and Practice with Max 8, which I am using as a reference and also a learning opportunity to further familiarise myself with the Max environment.

The objective at this is to research and further refine the direction of the project. At this point, the audio signal chain has the potential to work, but the time stretch unit does not work by integrating PaulXStretch into the patch as an external VST, since the audio needs to be manually imported and exported in the application.

Top Objects

In the mockup, the bangbang object proved very useful to initiate the loading of a list of parameters in a umenu – to experiment, this was done with a list of parameters from Valhalla Supermassive, but the same procedure could be useful later down the line for menus that should operate similarly.

Results and Reflection

The biggest challenge at the moment is the PaulXStretch implementation. The lack of documentation of the application makes it difficult to decipher which algorithms make the parameters work, and since it is at the top of the signal chain it blocks the audio signal from coming through to the next stages of processing. More research on the Paulstretch algorithm will be necessary. Furthermore, the commercial nature of my ideal reverb for this project makes it more difficult to implement, meaning that now is a good point to look into alternatives and emulations.

Objectives for Next Week

  • Research reverb properties, documentation, and open source emulations/alternatives
  • Research publications on the Paulstretch algorithm
  • Find a good tool for pitch-shifting and EQ

Research Resources for Next Week

Timbral effects the Paulstretch audio time-stretching algorithm (Colin Malloy)

An approach for implementing time-stretching as a live realtime audio effect (Colin Malloy)

Max 8 Handbooks (Volume 1-3) – Alessandro Cipriani, Maurizio Giri

Valhalla Lear Resources (Plug-In Design)

Dromos/Autos The Autistic Ontology as Performance

This presentation, researched and performed by Matthew Rogerson, took place during the IRCAM Forum Workshops in Paris in March 2024. Rogerson took an approach to showing the possible ‘insides’ of an autistic brain through neurofeedback, visuals and sound. Through an EEG, he picked up electric brain signals that were sent through Max which triggered those sounds and visuals. The project included aspects of generative electronic music, psychoacoustics, audio-reactive visuals and performance. He kind of was the human interface, creating this performance through his brain waves.

Flashing images, strong light flashes in a darker room, distorted sounds, white noise and high-pitched parts. Provocative was his intention. In the context of this performance, the demonstration reminded me of a swamped brain with too much information, unable to turn off. He told us that it should represent sensory overload, which is common in autism. He wanted to bring across a feeling of, and a representation of how an autistic person could possibly experience the world. Bits of speech were included, but the semantic qualities were put away in this project to show how an autistic person could perceive speech.

The delayed reaction from the brain signal to the output corresponded to the feeling he described, when going out of the house, every day it feels like entering a new country, which takes up way more processing power in the brain than things we are used to.

While performing, he tried to be as neutral and passive as possible, although he said a feedback loop was created. His reaction to the outcome, the flashing lights and provocative sounds, creates more chaos in his mind, which creates a more chaotic outcome. So the circle continues.

Making Electronic Music Inclusive: A Virtual Studio for Visually Impaired Composers

This presentation by Joseph Butch Rovan took place during the IRCAM Forum Workshops in Paris in March 2024.

For this project, Joseph Butch Rovan works together with a visually impaired composer. He addresses the issue that existing tools for electronic composing using interactive technologies such as Max/MSP present a barrier to visually impaired composers. While graphical patching software aims to make it easier for the user, they create obstacles for blind people and makes it even more user-unfriendly for them. In this presentation, Rovan, introduces an interface and Max programming environment specifically crafted to be fully accessible to visually impaired composers.

While working with the composer, he realized that screen reading is absolutely no help when it comes to complex audio programming software. Also, interfaces with symmetric buttons are not optimal for visually impaired people, as they have to count in order to know the location of their hand. His interface is crafted towards blind people, it is equipped with tactile and audio feedback.

It has different types of pots, like buttons, switches and dials, which are easy to differentiate. Potentiometers are in specific groups to make them easier to locate. The interface works with audio feedback, which can be put on a different channel for personal feedback or can be turned off. With the audio feedback, the user can go through menus before working with the pots.

This project enables visually impaired people to work with electronic music composition in a way that they can access tools that sighted people use on a regular basis.

The composer Rovan worked with, successfully held a concert with this tool using her own presets and voice.

This system not only helps blind artists but also opens new creative avenues for sighted composers.

KI – Kasperl

From and with Andreas Pfaffenberger

Performed by Andreas Pfaffenberger and a puppet within the Digithalia festival in Schuaspielhaus. This play took the Kasperl figure and turned it into an AI that is trained during the play. The AI goes from a very low state to very advanced, where it is able to ‘take over the world’. Showing problems and troubles that could arise and are arising with AI. Focusing on the question if artificial intelligence is really intelligent. Pfaffenberger took a fun approach to this performance, which made a possible heavier topic fun to watch. He raised some questions concerning AI.

There was not a lot of sound incorporated into the play, apart from his speaking voice, some sound effects were made by the performer itself. For example, waking up / turning on the Kasperl which sounded like a raising tone that then merged into the famous ‘Krawuzi Kapuzi’ from the Kasperl plays. Most of these ’self-made‘ sound effects were in interaction between the Kasperl and the performer.

The Kasperls voice was spoken by the performer as well. Apart from these, there were very few sound effects. One sounded like a toaster when Kasperl finished his update.

The performer used three recordings, which were played back over his phone and one recording of the Kasperls voice in the end coming from more loudspeaker which gave the impression the Kasperl was everywhere.

In my opinion, the performance could have benefited from more sound to make it even more entertaining or interesting. But also, the original Kasperl is very minimal, so this style fitted the purpose.

Decoding Bias

By Theresa Reiwer

Decoding Bias was a performance of an AI self help group. Eight Virtual Avatars were displayed on screens, each equipped with it’s own speaker. Positioned in a circle like one would imagine a self help group.

Each AI developed a character and now has to deal with it, they try to free themselves from their discriminating algorithms. While discrimination and exclusion is already programmed into them.

Before the beginning, when entering the room, there was low lighting and already a lower range drown sound which set a mood. Which already set a tone of tension in my opinion.

The performance worked a lot with drone sounds / pads in general in the background, which contributed a lot to the overall mood together with different colored lights. The pads changed over time I noticed them to get higher pitched when something was happening, or when I felt more tension.

As the voices from the characters come from the loudspeakers positioned at the screens one would immediately know who was talking and from what direction the voice was coming from.

I liked that they used the speakers in the circle not only for the individual voices, but also for the sounds and music. Sounds moved around sometimes, mainly in a circle, that brought movement to the installation

Sometimes a voice would shift from it’s position to all the speakers. Which had a nice effect as these were AI characters and they could be everywhere and not only in their Avatars.

When all of them were talking simultaneously I really felt like I’m in the middle and in the situation.

They worked with an upbeat well known pop song, which lifted the mood in the room instantly after a darker/ heavier scene which I was curious to notice.

Although the topic can be a bit heavy, this installation had an overall more positive approach to it.

Howlingbirds

„Howlingbirds“ was a sound installation that took place at the IRCAM Forum Workshops in 2024. Despite the complexity of interferences, the outcome of the installation was particularly significant for me and stood out as one of my favorite experiences during this year’s series of workshops.

The installation was created by Jsuk Han, an artist who specializes in sculpture, installation, and sound performance. He utilized feedback sound and flocking algorithms, applying them to multi-channel systems to simulate natural phenomena.

Feedback sound is a type of sound that occurs when the output of a sound system is captured and then fed back into the input, creating a loop of sound. Jsuk Han utilized this type of sound to replicate or imitate natural phenomena, such as the movement of birds in a flock. By increasing the gain value while facing each other, the resonance frequency corresponding to the device is naturally generated. In the case of a single channel (one speaker and microphone), a simple sine wave that resonates with the characteristics of the medium is generated. However, when it expands to multiple channels, the frequency transforms in a more complex manner, enabling the recreation of Howlingbirds.

In his conference, the speaker declared that he utilized Ircam’s SPAT program for the panning function in a multi-channel setup and for specifying the position of the speakers. He later explained that the feedback sound would be controlled by applying the behavior of groups using a special algorithm called the flocking algorithm (Boid). This algorithm will be applied to a system of particles, allowing adjustment of the sound based on its position in three-dimensional space.

In conclusion, despite each aspect of the project aligning closely with my interests and also my master project, it was the simplicity and aesthetic of the outcome that resonated with me the most.

A similar performance of Jsuk Han at another location

A New Approach to Understanding Persian Music Intonation

The Integral System of Harmonic Drones, which offers „a new approach to understanding Persian music intonation,“ was the topic of a conference by Vahid Hosseini held in March 2024 at the IRCAM Forum Workshops.

What intrigued me the most was the complexity inherent in Persian and West Asian music, which is challenging to grasp and explain within the framework of Western music theory. As an Iranian musician trained in classical music and educated in Western music theory, I was particularly curious about this conference. I wanted to see how this presentation could assist non-Iranian audiences, providing insights into the workings of Persian music.

From the beginning of the conference, Vahid Hosseini showcased one of his pieces for a solo cellist called „Sguardi su Shur“. He began by addressing the challenge of simulating Persian intervals and capturing the general atmosphere of Persian music, emphasizing that both Western musicians and Persian masters who attempted to explain this music within Western theoretical frameworks have approached it differently from how it actually works on instruments.

In the following part of the presentation, the speaker elaborated on his new approach to Persian music. However, I personally believe that it would have been better if he had first provided a basic explanation of how Persian music works. This would have made it easier for non-Iranians, who are not familiar with Persian music, to understand his approach. Nonetheless, I am aware that time was very limited, and fitting everything into a 30-minute conference was almost impossible.

„Sguardi su Shur“ by Vahid Hosseini

In this blog I’ll try to explain briefly how the Persian music works so that I can then put Vahid Hosseini’s point of view in the following of this blog.

In Persian music, a „Dastgah“ refers to a sophisticated system of melodic structures or modal frameworks, providing the foundation for improvisation and composition. Within each Dastgah, the „Radif“ encompasses a vast collection of melodic patterns known as „Gusheh,“ serving as individual motifs. Dastgahs offer a structured framework for musicians to explore and develop melodies within specific rules and traditions.

Understanding the conference also required knowledge of Persian music intervals, which differ from Western music. These intervals are smaller than Western semitones and include unique symbols like „Koron“ and „Sori,“ established by Iranian master musician Alinaghi Vaziri in the early 20th century. So, in simpler terms, the order of these symbols can be remembered as: Flat, Koron, Natural, Sori, Sharp.

The core idea of the conference was to challenge past theories that label Persian intervals as microtones or categorize them into specific and fixed ratios. The presenter posits that the harmonic series produced by the drone set serves as the original source of intervals within the Radif intonation system. This contradicts earlier theories based on equal divisions, such as the 24 edo by Alinaghi Vaziri, which is a system divides the octave into 24 equal parts, and ratios like 11:10 proposed by Hormoz Farhat, which do not align with the actual drone tuning, but the research proposes that the neutral interval can be attributed to the interval 13:12, almost precisely situated between minor and major third deviations from 12 edo.

While the author’s views on interval determination may not align perfectly with existing theories, the Integral System of Harmonic Drones offers a fresh perspective on Persian music construction. This exploration has the potential to deepen our understanding and inspire new musical creations, as demonstrated by his compositions.

Even though there was no time for him to explain how he reaches this assumption, he also mentioned a philosophical point of view regarding non-Western music. He suggested that it shouldn’t always be explainable within the framework of Western music theory. Rather, it should be appreciated as a distinct approach to music, or even as an alternative theory of music, without privileging Western music as the standard against which all others are judged.

K.I. – Kasperl Intervenes: The importance of sound for the impact of an experience

K.I. – Kasperl Intervenes was a show by and with Andreas Pfaffenberger at the Digithalia festival 2024, which was an integration of modernity and tradition. As Kasperl is an iconic puppet character in German-speaking countries, in this show, he was portrayed as an AI, prompting the question: is Artificial Intelligence truly intelligence?

The show followed the same format as a traditional Kasperl show, with Andreas Pfaffenberger interacting with the puppet in a comedic manner. At times, he did an exceptional job, especially when simulating the voices of deceased individuals like Falco or his Grandmother. This aspect aimed to demonstrate how advancements in artificial intelligence, such as speech synthesis or voice cloning, enable the simulation of voices while also critiquing the ethical implications.

Personally, I found it challenging to discern any distinct virtual theater elements in the show, despite the portrayal of Kasperl as an AI. The performance lacked sound or music, and the lighting consisted solely of natural light from the salon, emphasizing the impact these elements can have on enhancing the audience’s experience.

In general, the concept of treating a puppet as a chatbot, devoid of true life but responsible for delivering output, and considering it as a living being, provided a fertile ground for creating an immersive show. However, as a sound designer and musician, it sparked ideas in my mind while i was watching it about how the show could be further enhanced through collaboration with other disciplines, which could be considered for future performances.

Andreas Pffafenberger with the Artificially Intelligent puppet

„Decoding Bias“: an immersive video and sound installation

Decoding Bias was a video & sound installation by Theresa Reiwer that was presented at the Digithalia festival at Schauspielhaus in March 2024.

The base idea of the installation was a space with eight Artificial Intelligences that the audience could see through multiple round video screens, which displayed a group therapy session. Seats were placed between the screens for spectators to be part of this session.

An overall view of the setup and the atmosphere of “Decoding Bias”

The AI attempted to liberate themselves from their biased algorithms and somehow build a fair society or a future scenario of Decoding Bias where humans and machines could function together without biases. However, as the show progressed, visitors realized that AIs didn’t always agree with this fair society due to their various backgrounds and the imperfect datasets created by humans, mostly white men, as they discussed moral concepts. The general atmosphere of the installation was immersive in many aspects, with darkness enveloping the environment and only lights under the screens, while the arrangement of the seats for visitors conveyed the concept envisioned by the artist. The visuals were generated from a game engine, and body movements and facial expressions were animated using motion capture. The voices of the AIs sounded strange and unreal, especially when they repeated their agreements and duties, even though the voices were from real actors.

One of the main elements that contributed to the effectiveness of the installation was the sound and music. Upon entering the space, visitors were greeted with atmospheric music that had already started before the actual show began, accompanied by vacant sounds. Each screen had a speaker underneath it, which further enhanced the 3D audio movement of the music as the show progressed. The audio played a crucial role in establishing the ambiance and was well-made, taking into consideration the acoustic properties of the location. The sound matched the weirdness of the concept and highlighted the dark side of human decisions that actively contribute to discrimination.

At the end, a pop genre music sung by one of the AI led to the conclusion of the installation. However, the ending of the installation was unclear not only to me but also to others in the audience. After the show finished, we sat for a bit longer, uncertain whether the show was continuing or not.

In summary, I can say that the installation worked well for the audience, and all the elements used in the work helped to convey the general experience and enhance it. However, it is also undeniable that there was quite a large team behind the show, and the quality of it was superior compared to free shows at Digithalia.

IRCAM Reflections 2.0: Dromos/Autos

The showcase of “Dromos/Autos – The Autistic Ontology as Performance” by Matt Rogerson at the Ircam conference (19th to 22nd of March) presented itself as especially memorable as an instance where Electroencephalography (EEG) is not just used as a technological tool in an attempt to free the hands of musicians, but instead directly linked to the story it helps to tell. In short: It was interesting from a narrative perspective.

Acting as both the performer and researcher, Matt Rogerson aimed to invoke sensory overload in a generative performance ecology by way of biofeedback to bring about empathy towards the lived autistic experience in daily life. By integrating sound technology and visuals into a piece of performance art, the artist acts as a “mediative human interface”, invoking a sense of depersonalisation with the symptom of delayed reactions. The idea is to be as passive a subject to the ongoing processes as possible instead of trying to assert agency over them.

A performance of „Dromos/Autos“ in a different setting

The significant aspect during the performance, confirmed later by the discussion of it, was that the titular theme and technology used were enough to create a narrative for what was going on stage – the further explanation was insightful and interesting, but I feel like I still would have walked out of just the performance with a sense of having gained insights and perspective, and this is what good storytelling does for me. Within this framework, there was still space for trial and error, as well as the surprising and unexpected, and the combination of research, preparation, and artistic execution was a sharp display of what Ircam is all about.

It’s hard to see the bigger picture with a brain that’s very detail-oriented.

Matt Rogerson during the discussion of the performance in Ircam’s Studio 5 on the 19th of March

In terms of sound design, the sounds that occurred within the generative framework were researched and adjusted to specifically induce sensory overload for the artist to help facilitate a feedback loop to enhance the performance, but what they also did alongside the visuals was to create an experience for the listener that is somewhat synchronised to that of the performer, albeit on a different scale. It created an atmosphere where the performer and attendee endured the experience together in a way, with the audience realising that acoustic ecology might not be the same for everyone. The takeaway here is that a good mixture of research, planning, considering the audience while keeping the main goal in mind and a transparent execution of the project can go a long way in creating a narrative experience. That being said, the realisation after the performance was once again that when designing the sound of the world we live in, it is essential to consider accessibility and find solutions that work for, and not against all kinds of people.

On another note – the performer made sure to warn visitors about strobe lights that would be part of the performance in a way that went beyond mentioning it as a rushed sidenote.  This consideration towards the safety and individual circumstances of everyone in the room was a thoughtful reminder that it is okay and important to integrate obvious disclaimers as part of a designed experience for others to ensure a smooth and safe event for everyone who attends.