The SwissBERT model (Vamvas et al., SwissText 2023) extended by a Swiss German adapter that was trained on the character level.

Note: This model is experimental and can only be run with our codebase at https://github.com/ZurichNLP/swiss-german-text-encoders, since it uses a custom model architecture.

Training Data

For continued pre-training, we used the following two datasets of written Swiss German:

  1. SwissCrawl (Linder et al., LREC 2020), a collection of Swiss German web text (forum discussions, social media).
  2. A custom dataset of Swiss German tweets

In addition, we trained the model on an equal amount of Standard German data. We used news articles retrieved from Swissdox@LiRI.

License

Attribution-NonCommercial 4.0 International (CC BY-NC 4.0).

Citation

@inproceedings{vamvas-etal-2024-modular,
      title={Modular Adaptation of Multilingual Encoders to Written Swiss German Dialect},
      author={Jannis Vamvas and No{\"e}mi Aepli and Rico Sennrich},
      booktitle={First Workshop on Modular and Open Multilingual NLP},
      year={2024},
}
Downloads last month
6
Inference Examples
Inference API (serverless) has been turned off for this model.