ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2211.13613
22
18

Ham2Pose: Animating Sign Language Notation into Pose Sequences

24 November 2022
Rotem Shalev-Arkushin
Amit Moryossef
Ohad Fried
    SLR
ArXivPDFHTML
Abstract

Translating spoken languages into Sign languages is necessary for open communication between the hearing and hearing-impaired communities. To achieve this goal, we propose the first method for animating a text written in HamNoSys, a lexical Sign language notation, into signed pose sequences. As HamNoSys is universal, our proposed method offers a generic solution invariant to the target Sign language. Our method gradually generates pose predictions using transformer encoders that create meaningful representations of the text and poses while considering their spatial and temporal information. We use weak supervision for the training process and show that our method succeeds in learning from partial and inaccurate data. Additionally, we offer a new distance measurement for pose sequences, normalized Dynamic Time Warping (nDTW), based on DTW over normalized keypoints trajectories, and validate its correctness using AUTSL, a large-scale Sign language dataset. We show that it measures the distance between pose sequences more accurately than existing measurements and use it to assess the quality of our generated pose sequences. Code for the data pre-processing, the model, and the distance measurement is publicly released for future research.

View on arXiv
@article{shalev-arkushin2025_2211.13613,
  title={ Ham2Pose: Animating Sign Language Notation into Pose Sequences },
  author={ Rotem Shalev-Arkushin and Amit Moryossef and Ohad Fried },
  journal={arXiv preprint arXiv:2211.13613},
  year={ 2025 }
}
Comments on this paper