This paper addresses the gap in predicting turn-taking and backchannel actions in human-machine conversations using multi-modal signals (linguistic, acoustic, and visual). To overcome the limitation ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results