GMU:Max and the World/Leon Giacomo Goltermann: Difference between revisions

From Medien Wiki
No edit summary
No edit summary
 
(13 intermediate revisions by 2 users not shown)
Line 1: Line 1:
[[:File:homework_1.maxpat]]


[[:File:homework2_leon.maxpat]]
[[File:Bildschirmfoto 2021-05-12 um 18.08.24.png|400px]]


Prototype to test my Installation ideas. (it is reduced to triggering sounds with buttons insted of tracking).


'''New version with instruction in the patch.
I built a patch that uses computer-vision to control sound-plaback with head-gestures.  
[[:File:prototyp instalationsinterface.maxpat]]
Here is the patch and a video that explains how to use it and what you can do with it.
'''
here you can find soundfiles to try the patch or you use your own sound/max-library
https://cloud.uni-weimar.de/s/znnFMEzCJGEcYBD


the purpose is documented in the video https://cloud.uni-weimar.de/s/ig53DAm3naP7Rfz
[[:File:facetracking_sampler_leon_g.maxpat]]


{{#ev:youtube|mef2j1WzQoI}}


As an evaluated project for the this course i want to produce a prototype that shows the possibilities of the installation - Lesekomposition -  (wt) i am building.
It should provide the connection of sounds and words. The sounds need to be triggered (by knobs or with sensor data). It should also be possible that sounds are changed/manipulated by the combination of word/triggers; there should be a library of sounds for each trigger and different presets for selecting different lists.


[[:File:versuch sprache rythmus.maxpat]]
 
The patch was developed in the process of building an interactive installation experimenting with the extension of language in the combination of sound and language.
 
I realized the Prototype - compo - which provides an instrument to create future interactive compositions between words and music.
 
It is documented over here: https://wwws.uni-weimar.de/kunst-und-gestaltung/wiki/GMU:Artists_Lab_IV/Leon_Goltermann

Latest revision as of 15:12, 14 May 2021

Bildschirmfoto 2021-05-12 um 18.08.24.png


I built a patch that uses computer-vision to control sound-plaback with head-gestures. Here is the patch and a video that explains how to use it and what you can do with it.

File:facetracking_sampler_leon_g.maxpat


The patch was developed in the process of building an interactive installation experimenting with the extension of language in the combination of sound and language.

I realized the Prototype - compo - which provides an instrument to create future interactive compositions between words and music.

It is documented over here: https://wwws.uni-weimar.de/kunst-und-gestaltung/wiki/GMU:Artists_Lab_IV/Leon_Goltermann