學術產出-Proceedings

Article View/Open

Publication Export

Google ScholarTM

政大圖書館

Citation Infomation

題名 Audio-driven facial landmark generation in violin performance using 3DCNN network with self attention model
作者 劉昭麟
Liu, Chao-Lin;Lin, Ting-Wei;Su, Li
貢獻者 資訊系
關鍵詞 music to face generation; facial landmarks generation; music-face dataset; 3DCNN; self-attention
日期 2023-06
上傳時間 30-Nov-2023 11:26:28 (UTC+8)
摘要 In a music scenario, both auditory and visual elements are essential to achieve an outstanding performance. Recent research has focused on the generation of body movements or fingering from audio in music performance. The audio-driven face generation technique in music performance is still deficient. In this paper, we compile a violin soundtrack and facial expression dataset (VSFE) for modeling facial expressions in violin performance. To our knowledge, this is the first dataset mapping the relationship between violin performance audio and musicians’ facial expressions. We then propose a 3DCNN network with self-attention and residual blocks for audio-driven facial expression generation. In the experiments, we compare our methods with three baselines on talking face generation.
關聯 Proceedings of the 2023 IEEE International Conference on Acoustics, Speech and Signal Processing, IEEE, pp.1-5
資料類型 conference
DOI https://doi.org/10.1109/ICASSP49357.2023.10096358
dc.contributor 資訊系
dc.creator (作者) 劉昭麟
dc.creator (作者) Liu, Chao-Lin;Lin, Ting-Wei;Su, Li
dc.date (日期) 2023-06
dc.date.accessioned 30-Nov-2023 11:26:28 (UTC+8)-
dc.date.available 30-Nov-2023 11:26:28 (UTC+8)-
dc.date.issued (上傳時間) 30-Nov-2023 11:26:28 (UTC+8)-
dc.identifier.uri (URI) https://nccur.lib.nccu.edu.tw/handle/140.119/148299-
dc.description.abstract (摘要) In a music scenario, both auditory and visual elements are essential to achieve an outstanding performance. Recent research has focused on the generation of body movements or fingering from audio in music performance. The audio-driven face generation technique in music performance is still deficient. In this paper, we compile a violin soundtrack and facial expression dataset (VSFE) for modeling facial expressions in violin performance. To our knowledge, this is the first dataset mapping the relationship between violin performance audio and musicians’ facial expressions. We then propose a 3DCNN network with self-attention and residual blocks for audio-driven facial expression generation. In the experiments, we compare our methods with three baselines on talking face generation.
dc.format.extent 113 bytes-
dc.format.mimetype text/html-
dc.relation (關聯) Proceedings of the 2023 IEEE International Conference on Acoustics, Speech and Signal Processing, IEEE, pp.1-5
dc.subject (關鍵詞) music to face generation; facial landmarks generation; music-face dataset; 3DCNN; self-attention
dc.title (題名) Audio-driven facial landmark generation in violin performance using 3DCNN network with self attention model
dc.type (資料類型) conference
dc.identifier.doi (DOI) 10.1109/ICASSP49357.2023.10096358
dc.doi.uri (DOI) https://doi.org/10.1109/ICASSP49357.2023.10096358