Dmitry Ryumin
DmitryRyumin
AI & ML interests
Machine Learning and Applications, Multi-Modal Understanding
Recent Activity
reacted
to
their
post
with ๐ค
2 days ago
๐๐๐ New Research Alert - ICCV 2025 (Oral)! ๐๐ค๐
๐ Title: Understanding Co-speech Gestures in-the-wild ๐
๐ Description: JEGAL is a tri-modal model that learns from gestures, speech and text simultaneously, enabling devices to interpret co-speech gestures in the wild.
๐ฅ Authors: @sindhuhegde, K R Prajwal, Taein Kwon, and Andrew Zisserman
๐
Conference: ICCV, 19 โ 23 Oct, 2025 | Honolulu, Hawai'i, USA ๐บ๐ธ
๐ Paper: https://huggingface.co/papers/2503.22668
๐ Web Page: https://www.robots.ox.ac.uk/~vgg/research/jegal
๐ Repository: https://github.com/Sindhu-Hegde/jegal
๐บ Video: https://www.youtube.com/watch?v=TYFOLKfM-rM
๐ ICCV-2023-25-Papers: https://github.com/DmitryRyumin/ICCV-2023-25-Papers
๐ Added to the Human Modeling Section: https://github.com/DmitryRyumin/ICCV-2023-25-Papers/blob/main/sections/2025/main/human-modeling.md
๐ More Papers: more cutting-edge research presented at other conferences in the https://huggingface.co/spaces/DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin
๐ Keywords: #CoSpeechGestures #GestureUnderstanding #TriModalRepresentation #MultimodalLearning #AI #ICCV2025 #ResearchHighlight
reacted
to
their
post
with ๐ฅ
2 days ago
๐๐๐ New Research Alert - ICCV 2025 (Oral)! ๐๐ค๐
๐ Title: Understanding Co-speech Gestures in-the-wild ๐
๐ Description: JEGAL is a tri-modal model that learns from gestures, speech and text simultaneously, enabling devices to interpret co-speech gestures in the wild.
๐ฅ Authors: @sindhuhegde, K R Prajwal, Taein Kwon, and Andrew Zisserman
๐
Conference: ICCV, 19 โ 23 Oct, 2025 | Honolulu, Hawai'i, USA ๐บ๐ธ
๐ Paper: https://huggingface.co/papers/2503.22668
๐ Web Page: https://www.robots.ox.ac.uk/~vgg/research/jegal
๐ Repository: https://github.com/Sindhu-Hegde/jegal
๐บ Video: https://www.youtube.com/watch?v=TYFOLKfM-rM
๐ ICCV-2023-25-Papers: https://github.com/DmitryRyumin/ICCV-2023-25-Papers
๐ Added to the Human Modeling Section: https://github.com/DmitryRyumin/ICCV-2023-25-Papers/blob/main/sections/2025/main/human-modeling.md
๐ More Papers: more cutting-edge research presented at other conferences in the https://huggingface.co/spaces/DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin
๐ Keywords: #CoSpeechGestures #GestureUnderstanding #TriModalRepresentation #MultimodalLearning #AI #ICCV2025 #ResearchHighlight
posted
an
update
2 days ago
๐๐๐ New Research Alert - ICCV 2025 (Oral)! ๐๐ค๐
๐ Title: Understanding Co-speech Gestures in-the-wild ๐
๐ Description: JEGAL is a tri-modal model that learns from gestures, speech and text simultaneously, enabling devices to interpret co-speech gestures in the wild.
๐ฅ Authors: @sindhuhegde, K R Prajwal, Taein Kwon, and Andrew Zisserman
๐
Conference: ICCV, 19 โ 23 Oct, 2025 | Honolulu, Hawai'i, USA ๐บ๐ธ
๐ Paper: https://huggingface.co/papers/2503.22668
๐ Web Page: https://www.robots.ox.ac.uk/~vgg/research/jegal
๐ Repository: https://github.com/Sindhu-Hegde/jegal
๐บ Video: https://www.youtube.com/watch?v=TYFOLKfM-rM
๐ ICCV-2023-25-Papers: https://github.com/DmitryRyumin/ICCV-2023-25-Papers
๐ Added to the Human Modeling Section: https://github.com/DmitryRyumin/ICCV-2023-25-Papers/blob/main/sections/2025/main/human-modeling.md
๐ More Papers: more cutting-edge research presented at other conferences in the https://huggingface.co/spaces/DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin
๐ Keywords: #CoSpeechGestures #GestureUnderstanding #TriModalRepresentation #MultimodalLearning #AI #ICCV2025 #ResearchHighlight