musique contemporaine

Ircam - articles scientifiques notice originale

Spatial rendering of audio-visual synthetic speech use for immersive environments

Type

text
 

Genre(s)

article
 

Forme(s)

document imprimé
 

Cette ressource est disponible chez l'organisme suivant : Ircam - Centre Pompidou

Identification

Titre

Spatial rendering of audio-visual synthetic speech use for immersive environments
 

Nom(s)

Noisternig, Markus (auteur)
 
Katz, Brian F. G. (auteur)
 
D'Alessandro, Christophe (auteur)
 

Publication

Paris, France , 2008
 

Description

Sujet(s)

Perception of voice and talker characteristics   computer simulation of acoustics in enclosures
 

Résumé

Synthetic speech is usually delivered as a mono audio signal. In this pro ject, audio-visual speech synthesis is attributed to a virtual agent moving in a virtual 3-dimensional scene. More realistic acoustic rendering is achieved by taking into account the position of the agent in the scene, the acoustics of the room depicted in the scene, and the orientation of the virtual character's head relative. 3D phoneme dependant radiation patterns have been measured for two speakers and a singer. These data are integrated into a Text-To-Speech system using a phoneme to directivity pattern transcription module which also includes a phoneme to viseme model for the agent. In addition to the effects related to agent's head orientation for the direct sound, a room acoustics model allows for realistic rendering of the room effect as well as the apparent distance as depicted in the virtual scene. Real-time synthesis is implemented in a 3D audio rendering system.
 

Note(s)

Contribution au colloque ou congrès : 155th ASA, 5th Forum Austicum, and 2nd ASA-EAA Joint Conference (Acoustics'08)
 

Localisation

Envoyer la notice

Bookmark and Share 
 

Identifiant OAI

 

Date de la notice

2014-11-18 01:00:00
 

Identifiant portail

 

Contact