• Keine Ergebnisse gefunden

Speech animation using electromagnetic articulography as motion capture data

N/A
N/A
Protected

Academic year: 2022

Aktie "Speech animation using electromagnetic articulography as motion capture data"

Copied!
1
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)Speech animation using electromagnetic articulography as motion capture data Ingmar Steiner. Korin Richmond. Slim Ouni. ingmar.steiner@dfki.de. korin@cstr.ed.ac.uk. slim.ouni@loria.fr. NIVER. S. R. E. G. H. Y. O F. Overview. IT. TH E. U. D I U N B. EMA data as motion capture. Electromagnetic articulography (EMA) captures the position and orientation of a number of markers, attached to the articulators, during speech. As such, it performs the same function for speech that conventional, marker-based motion capture does for fullbody movements acquired with optical modalities. We present an approach to processing EMA data from a motion-capture perspective and applying it to the visualization of an existing multimodal corpus of articulatory data,a creating a kinematic 3D model of the articulators by adapting a conventional motion capture based animation paradigm. Such an animated model can then be easily integrated into multimedia applications as an animation asset, allowing the visualization of speech production in an intuitive and accessible manner. a http://mngu0.org/. HIERARCHY ROOT T3 { OFFSET −0.08384774 −0.68610632 0.72265321 CHANNELS 6 XPosition YPosition ZPosition XRotation YRotation ZRotation End Site { OFFSET 0.00000000 −1.00000000 0.00000000 } } ROOT upperlip { OFFSET −0.94566417 0.29789653 −0.13029577 CHANNELS 6 XPosition YPosition ZPosition XRotation YRotation ZRotation End Site { OFFSET 0.00000000 −1.00000000 0.00000000 } } MOTION Frames : 724 Frame Time : 0.005 6.78959465 12.27275467 1.20037365 −8.13653286 56.68709013 −1.78241577 . . . −6.08323193 12.96489048 0.38591787 19.70638666 −11.45516554 −52.56656678 . . . 0.17483102 0.97839558 −3.00247645 3.93864096 24.40054857 −51.69048915 . . . 0.05262496 −0.48740909 −2.57859731 56.49315374 9.14393027 −2.77820826 . . . 0.00545616 −0.00614901 5.71714687 56.73992164 1.99134613 7.70857967 . . . −0.00336362 −0.02571771 0.00576410 −57.27868696 −1.34396448 0.38998995 . . . 0.25838542 1.55094087 −0.98620248 7.96369151 −30.90922537 −47.58157029 . . . 0.10555566 3.21496391 0.25413591 −8.54249967 −56.29318629 −6.39601923 . . . 0.11248340 5.05562544 −0.04919426 −4.80412156 −39.31099667 41.40497899 . . . 0.09581324 −1.06100082 −0.19431658 −54.18256563 17.06821414 −7.46539762. Fragment of one EMA sweep from mngu0 database in Biovision Hierarchy format. EMA coils are rendered over a 250 ms window (frame step = 3); lips, incisors at left, tongue coils 1 to 3 at right.. Articulatory model The tongue is segmented from a volumetric magnetic resonance imaging (MRI) scan and retopologized into a mesh. The tongue mesh can be deformed using spline inverse kinematics (IK); the spline’s control points are modified by the EMA coils. Dental scans are registered into the same space and added to the rig.. Volume rendering of raw MRI data. Voxel-tesselated isosurface from MRI. Retopologized tongue mesh. Tongue model rigged for spline IK, with mandible. Animation. For speech animation, the EMA data drives the animation rig. The ref and jaw coils control the maxilla and mandible, respectively; the tongue coils move the IK control points, which in turn deform the tongue mesh..

(2)

Referenzen

ÄHNLICHE DOKUMENTE

1) Building Block 1: Global Topology 1riformation The topological information which is used for indoor navigation models the relations between rooms and ways and

According to Jacobson, business processes could only change once the current problems are identified. The new or changed processes and their differences from

The methods have been applied with two different types of input data: (1) binary forest presence/absence; (2) the percentage of forest cover within a grid cell.. Both types of data

The maximum error in the elevation angle analysis due to soft tissue displacement and motion capture data processing typically remained below 1°, see Error! Reference source

In applying this, we distinguish between (visual) analysis for exploration and validation. Especially the latter turns out to be hard in some cases, indicated by a qualitative

We identified the following problems we aim at supporting with MotionExplorer: 1.) As motion data collections are typically large, an overview is needed. To our knowledge, visual

Like any user interface and interaction design process, designing information visualization should follow an end-user-centered approach, not only limited to pragmatic qualities

An algorithm developed at the Faculty of Mathematics of the University of Vienna enables the signals received by WIM sensors to be assigned correctly to individual axles and