Audio-Visual Model for Generating Eating Sounds Using Food ASMR Videos
We present an audio-visual model for generating food texture sounds from silent eating videos. We designed a deep network-based model that takes the visual features of the detected faces as input and outputs a magnitude spectrogram that aligns with the visual streams. Generating raw waveform samples...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9388653/ |
id |
doaj-799d96852ff649f4b762e772e76d499e |
---|---|
record_format |
Article |
spelling |
doaj-799d96852ff649f4b762e772e76d499e2021-04-05T17:38:25ZengIEEEIEEE Access2169-35362021-01-019501065011110.1109/ACCESS.2021.30692679388653Audio-Visual Model for Generating Eating Sounds Using Food ASMR VideosKodai Uchiyama0Kazuhiko Kawamoto1https://orcid.org/0000-0003-3701-1961Graduate School of Science and Engineering, Chiba University, Chiba, JapanGraduate School of Engineering, Chiba University, Chiba, JapanWe present an audio-visual model for generating food texture sounds from silent eating videos. We designed a deep network-based model that takes the visual features of the detected faces as input and outputs a magnitude spectrogram that aligns with the visual streams. Generating raw waveform samples directly from a given input visual stream is challenging; in this study, we used the Griffin-Lim algorithm for phase recovery from the predicted magnitude to generate raw waveform samples using inverse short-time Fourier transform. Additionally, we produced waveforms from these magnitude spectrograms using an example-based synthesis procedure. To train the model, we created a dataset containing several food autonomous sensory meridian response videos. We evaluated our model on this dataset and found that the predicted sound features exhibit appropriate temporal synchronization with the visual inputs. Our subjective evaluation experiments demonstrated that the predicted sounds are considerably realistic to fool participants in a “real” or “fake” psychophysical experiment.https://ieeexplore.ieee.org/document/9388653/Multi-modal deep neural networkautonomous sensory meridian responseeating sound generation |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Kodai Uchiyama Kazuhiko Kawamoto |
spellingShingle |
Kodai Uchiyama Kazuhiko Kawamoto Audio-Visual Model for Generating Eating Sounds Using Food ASMR Videos IEEE Access Multi-modal deep neural network autonomous sensory meridian response eating sound generation |
author_facet |
Kodai Uchiyama Kazuhiko Kawamoto |
author_sort |
Kodai Uchiyama |
title |
Audio-Visual Model for Generating Eating Sounds Using Food ASMR Videos |
title_short |
Audio-Visual Model for Generating Eating Sounds Using Food ASMR Videos |
title_full |
Audio-Visual Model for Generating Eating Sounds Using Food ASMR Videos |
title_fullStr |
Audio-Visual Model for Generating Eating Sounds Using Food ASMR Videos |
title_full_unstemmed |
Audio-Visual Model for Generating Eating Sounds Using Food ASMR Videos |
title_sort |
audio-visual model for generating eating sounds using food asmr videos |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2021-01-01 |
description |
We present an audio-visual model for generating food texture sounds from silent eating videos. We designed a deep network-based model that takes the visual features of the detected faces as input and outputs a magnitude spectrogram that aligns with the visual streams. Generating raw waveform samples directly from a given input visual stream is challenging; in this study, we used the Griffin-Lim algorithm for phase recovery from the predicted magnitude to generate raw waveform samples using inverse short-time Fourier transform. Additionally, we produced waveforms from these magnitude spectrograms using an example-based synthesis procedure. To train the model, we created a dataset containing several food autonomous sensory meridian response videos. We evaluated our model on this dataset and found that the predicted sound features exhibit appropriate temporal synchronization with the visual inputs. Our subjective evaluation experiments demonstrated that the predicted sounds are considerably realistic to fool participants in a “real” or “fake” psychophysical experiment. |
topic |
Multi-modal deep neural network autonomous sensory meridian response eating sound generation |
url |
https://ieeexplore.ieee.org/document/9388653/ |
work_keys_str_mv |
AT kodaiuchiyama audiovisualmodelforgeneratingeatingsoundsusingfoodasmrvideos AT kazuhikokawamoto audiovisualmodelforgeneratingeatingsoundsusingfoodasmrvideos |
_version_ |
1721539180146720768 |