This is a promo video for the next Lovethechaos release, a cassette in which I’ve been working during the last year. The video was generated from the sound, both things are the same files, so what you see is the same that what you hear.

More about the idea and the process from the promo sheet:

The 186 tracks of this cassette are 186 photographs of cassettes opened as sound. It is not data visualisation or a random connection between image and sound, but a series of JPG files exported as AIFF. That is, a direct digital translation based on the old idea that colour (image) and sound (music) share some similarities that allow the translation of one media into the other. Even it is not possible to save a JPG as AIFF, it is possible to save it as RAW, a digital format compatible with both image and sound editing software. This means that any digital image can be automatically transformed into sound, and vice versa, without carrying out processes of sonification, data visualisation or any other kind of random or arbitrary transformation.

The theory that states than colour/light and sound are intrinsically related comes from Pythagoras and Aristotle, who linked each colour to a specific music interval. Pythagoras linked each creative sphere to a tone, a harmonic interval and a colour, among other things, while Aristotle stated that:

“…these colours [viz. all those based on numerical ratios] as analogous to the sounds that enter into music, and suppose that those involving simple numerical ratios, like the concords in music, may be those generally regarded as most agreeable; as, for example, purple, crimson, and some few such colours, their fewness being due to the same causes which render the concords few. The other compound colours may be those which are not based on numbers. Or it may be that, while all colours whatever [except black and white] are based on numbers, some are regular in this respect, others irregular; and that the latter [though now supposed to be all based on numbers], whenever they are not pure, owe this character to a corresponding impurity in [the arrangement of] their numerical ratios.” On Sense and the Sensible, Aristotle.

These theories were expanded and rectified by many subsequent thinkers, scientists and artists, from Newton — who divided the light spectrum into seven colours, one for each note of the musical scale – to Arcimboldo — who tried to match painting with music.

Most of these theories are based on the fact that both light and sound are types of waves, so it is possible to establish direct analogies based on frequency data. Nevertheless, the range of frequencies in which light and sound coexist is quite limited.

The first practical audiovisual implementations of these theories emerged on the 18th century with devices as the Ocular Harpsichord (1725) by Louis Bertrand Castel, which later derived in instruments such as Clavilux (1919) by Thomas Wilfred and Sarabet (1922) by Mary Hallock-Greenewalt. In any case, the relationships between light and sound established in these instruments were arbitrary, not direct translations. The closest thing that we can find to a direct translation from image to sound are the optical sound experiments of experimental filmmakers such as Norman McLaren, Lis Rhodes and Guy Sherwin.

Before digital sound and analogous sound, motion pictures had an optical soundtrack, that is, the soundtrack was printed directly on the margin of the photographic film. Some artists took advantage of this fact to generate the soundtrack from the film images, instead of using a standard soundtrack, which lead to experiments as Synchromy (1971) by Normal McLaren, Dresden Dynamo (1971) by Lis Rhodes and Railings (1977) by Guy Sherwin, three films in which what you see is exactly the same that what you hear.

Such contemporary experiments were not only made by visual artists, some engineers and musicians built photoelectric instruments, as the ANS Synthesiser (1937-1957) and the Oramics (1957), that synthesised sounds from drawings. Other similar instrument from the digital era is the UPIC by Iannis Xenakis, a compositional tool based on a tablet in which the user draws waveforms and sound envelopes.

The technique used to generate this cassette is based on the same idea that optical sound and visual synthesisers, but it belongs to a recent digital genre called data bending. Data bending (or databending) are a series of methods that allow the transformation of a type of file into another type of file (for example, a sound file into an image file). Even data bending is related with glitch aesthetics, data bending errors are always provoked, they do not happen by chance. However, the result is always unexpected — since the transcodification is automatic or semiautomatic it is impossible to predict its consequences.

2 thoughts on “LTC016”

Leave a Reply

Your email address will not be published. Required fields are marked *