Hackathon and performance on Generative Audiovisuals
Performance: 6/2/2015, Goldsmiths, University of London, Cinema
About the projects:
In ABP, a PureData patch creates music and sends data to the Cinder visuals software via OSC.
Several visual parameters like color, alpha, zoom, repetition of objects, tempo are given by the PureData patch.
Project link: https://github.com/AVUIs/AdamBrucePiotr
drawSynth consists of a GUI to control sound and image. Users can draw shapes and select colours. By doing this, users control the synthesis engine. Each pair of vertices of a shape will control an FM oscillator – one vertex for carrier and the other for the modulator. The project is built with openFrameworks for graphics and interaction, and Max/MSP for sound, using FM synthesis. OSC is used for communication between both.
Project link: https://github.com/AVUIs/drawSynth
Esoterion Universe consists of an empty 3D space that can be filled with planet-like audiovisual objects. The objects can be manipulated and can be given different appearances and sounds. Users can navigate in space and the audiovisual outcome is influenced by that navigation. Generic, media neutral terms such as warmth, sharpness, size and oscillation are used to characterise and connect sound and visuals. A semantic approach was chosen for this connection instead of a one-to-one parameter mapping. The GUI consists of sliders distributed concentrically, in the shape of a star graph, embedded in the centre of the object. It integrates aesthetically with the objects. openFrameworks with openGL is used for graphics and interaction, and Max/MSP for sound. OSC is used for communication.
Project link: https://github.com/AVUIs/EsoterionUniverse
GS.avi is a gestural instrument that generates continuous spatial visualizations and music from the input of a performer. The features extracted form a performer’s gesture defines the color, position, form and orientation of a 3-demensional Delaunay mesh — its composite triangles, vertices, edges and walk. The music, composed using granular synthesis, is generated from features extracted from the mesh — its colors, strokes, position, orientation and patterns. The project was created using Processing and MAX/MSP. OSC is used to communicate between the two.
Project link: https://github.com/AVUIs/GS.avi
Modulant allows for the creation of images and their sonification. The present implementation is built upon image-importing and freehand-drawing modules that may be used to create arbitrary visual scenes, with more constrained functional and typographical modules in development. The audio engine is inspired by a 1940’s synthesizer, the ANS, that scans across images. In this scanning, one axis is time and the other axis is frequency. Modulant thus becomes a graphical space to be explored sonically and vice-versa. The project is built with Processing for graphics and interaction, and Ruby with Puredata for sound.
Project link: https://github.com/AVUIs/Modulant
Projects developed during the Hackathon on Interactive Computer-Generated Audiovisuals (6-7 December 2014, Goldsmiths, University of London).