ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2103.15963
13
6

Contextual Text Embeddings for Twi

29 March 2021
P. Azunre
Salomey Osei
S. Addo
Lawrence Asamoah Adu-Gyamfi
Stephen E. Moore
Bernard Adabankah
Bernard Opoku
Clara Asare-Nyarko
S. Nyarko
Cynthia Amoaba
Esther Dansoa Appiah
Felix Akwerh
Richard Nii Lante Lawson
Joel Budu
E. Debrah
N. Boateng
Wisdom Ofori
Edwin Buabeng-Munkoh
F. Adjei
Isaac. K. E. Ampomah
Joseph Otoo.
R. Borkor
Standylove Birago Mensah
Lucien Mensah
Mark Amoako Marcel
A. Amponsah
J. B. Hayfron-Acquah
ArXivPDFHTML
Abstract

Transformer-based language models have been changing the modern Natural Language Processing (NLP) landscape for high-resource languages such as English, Chinese, Russian, etc. However, this technology does not yet exist for any Ghanaian language. In this paper, we introduce the first of such models for Twi or Akan, the most widely spoken Ghanaian language. The specific contribution of this research work is the development of several pretrained transformer language models for the Akuapem and Asante dialects of Twi, paving the way for advances in application areas such as Named Entity Recognition (NER), Neural Machine Translation (NMT), Sentiment Analysis (SA) and Part-of-Speech (POS) tagging. Specifically, we introduce four different flavours of ABENA -- A BERT model Now in Akan that is fine-tuned on a set of Akan corpora, and BAKO - BERT with Akan Knowledge only, which is trained from scratch. We open-source the model through the Hugging Face model hub and demonstrate its use via a simple sentiment classification example.

View on arXiv
Comments on this paper