Latent Embeddings
Research output: Creative and Literary Works in Non-textual Form (RGC: 42, 43, 44, 46, 47) › 43_Film, video
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Publication status | Published - 12 Nov 2019 |
Exhibition
Title | HKACT! Act 9 WYSIWYG Jeffrey Shaw Solo Exhibition |
---|---|
Location | Osage Gallery |
Place | Hong Kong |
Period | 13 November 2019 - 21 June 2020 |
Link(s)
Document Link | Links
|
---|---|
Permanent Link | https://scholars.cityu.edu.hk/en/publications/publication(42df07d8-b679-448a-9f33-6c5863d3fb85).html |
Abstract
This project is a collaboration of artists Jeffrey Shaw and Hector Rodriguez, with coding by Sam Chan and technical advisor Mike Wong. It premiered at the Jeffrey Shaw retrospective exhibition entitled WYSIWYG in the Osage Gallery, Hong Kong.
This work uses unsupervised machine learning techniques to revisit the animated film Continuous Sound and Image Moments, made in 1966 by Jeffrey Shaw, Willem Breuker, and Tjebbe van Tijen. The original film consisted of a sequence of hand drawings, each shown for only a few seconds.n Latent Embeddings, digitized versions of those drawings are processed by a Vector Quantized Variational Autoeconder (VQ-VAE), an algorithm that uses variational inference techniques to construct a generative model. New images are then produced by exploring the latent space of the model.
These images resemble the hand-drawn originals to varying degrees. The artists control the extent of resemblance by constraining the algorithm either to remain close to the training dataset or to venture into more remote areas of the latent space, fluctuating between recognition and surprise, between memory and discovery.
In the physical installation, a set of computer-generated images are displayed on a monitor with a throttle bar that enables users to navigate them at any speed.
This work uses unsupervised machine learning techniques to revisit the animated film Continuous Sound and Image Moments, made in 1966 by Jeffrey Shaw, Willem Breuker, and Tjebbe van Tijen. The original film consisted of a sequence of hand drawings, each shown for only a few seconds.n Latent Embeddings, digitized versions of those drawings are processed by a Vector Quantized Variational Autoeconder (VQ-VAE), an algorithm that uses variational inference techniques to construct a generative model. New images are then produced by exploring the latent space of the model.
These images resemble the hand-drawn originals to varying degrees. The artists control the extent of resemblance by constraining the algorithm either to remain close to the training dataset or to venture into more remote areas of the latent space, fluctuating between recognition and surprise, between memory and discovery.
In the physical installation, a set of computer-generated images are displayed on a monitor with a throttle bar that enables users to navigate them at any speed.
Bibliographic Note
Record validation is based on the information provided by the researcher(s) concerned. Suggestions for further amendments or supplementary information can be sent to lbscholars@cityu.edu.hk.
Citation Format(s)
Latent Embeddings. Shaw, Jeffrey (Artist); Rodriguez, Hector (Artist); Chan, Sam (Developer) et al.. 2019. Event details: HKACT! Act 9 WYSIWYG Jeffrey Shaw Solo Exhibition, Osage Gallery, Hong Kong.
Research output: Creative and Literary Works in Non-textual Form (RGC: 42, 43, 44, 46, 47) › 43_Film, video