AI Dynamics

Global AI News Aggregator

Multimodal AI improves speech by analyzing gaze and vocal patterns

One other thing I noticed is that people still had to speak without stopping. Notice how they never take a breath or pause? Multimodal can address this by looking at where someone is gazing and whether they're using umms and uhhs. https://
angelicalim.com/papers/humanoi
ds2017_paper.pdf

→ View original post on X — @petitegeek,

Commentaires

Leave a Reply

Your email address will not be published. Required fields are marked *