ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.08173
22
0

Semi-Supervised Spoken Language Glossification

12 June 2024
Huijie Yao
Wengang Zhou
Hao Zhou
Houqiang Li
ArXivPDFHTML
Abstract

Spoken language glossification (SLG) aims to translate the spoken language text into the sign language gloss, i.e., a written record of sign language. In this work, we present a framework named SSSemi-SSSupervised SSSpoken LLLanguage GGGlossification (S3S^3S3LG) for SLG. To tackle the bottleneck of limited parallel data in SLG, our S3S^3S3LG incorporates large-scale monolingual spoken language text into SLG training. The proposed framework follows the self-training structure that iteratively annotates and learns from pseudo labels. Considering the lexical similarity and syntactic difference between sign language and spoken language, our S3S^3S3LG adopts both the rule-based heuristic and model-based approach for auto-annotation. During training, we randomly mix these complementary synthetic datasets and mark their differences with a special token. As the synthetic data may be less quality, the S3S^3S3LG further leverages consistency regularization to reduce the negative impact of noise in the synthetic data. Extensive experiments are conducted on public benchmarks to demonstrate the effectiveness of the S3S^3S3LG. Our code is available at \url{https://github.com/yaohj11/S3LG}.

View on arXiv
Comments on this paper