Abstract
Enabling online virtual reality (VR) users to dance and move in a way that mirrors the real-world necessitates improvements in the accuracy of predicting human motion sequences paving way for an immersive and connected experience. However, the drawbacks of latency in networked motion tracking present a critical detriment in creating a sense of complete engagement, requiring prediction for online synchronization of remote motions. To address this challenge, we propose a novel approach that leverages a synthetically generated dataset based on supervised foot anchor placement timings of rhythmic motions to ensure periodicity resulting in reduced prediction error. Specifically, our model compromises a discrete cosine transform (DCT) to encode motion, refine high frequencies and smooth motion sequences and prevent jittery motions. We introduce a feed-forward attention mechanism to learn based on dual-window pairs of 3D key points pose histories to predict future motions. Quantitative and qualitative experiments validating on the Human3.6m dataset result in observed improvements in the MPJPE evaluation metrics protocol compared with prior state-of-the-art.
| Original language | English |
|---|---|
| Title of host publication | The 42nd Eurographics UK Conference on Computer Graphics Visual Computing Conference, CVGC |
| Publisher | The Eurographics Association |
| ISBN (Print) | 9783038682493 |
| DOIs | |
| Publication status | Published - 2024 |
| Event | 42nd Computer Graphics & Visual Computing Conference 2024 - City University of London, London, United Kingdom Duration: 12 Sept 2024 → 13 Sept 2024 https://cgvc.org.uk/CGVC2024/ |
Conference
| Conference | 42nd Computer Graphics & Visual Computing Conference 2024 |
|---|---|
| Abbreviated title | CGVC 2024 |
| Country/Territory | United Kingdom |
| City | London |
| Period | 12/09/24 → 13/09/24 |
| Internet address |