IsROBERTa

The First Icelandic Large Transformer Language Model

isroberta-neurocode

Technology

PythonPytorchONNXWandBPoetryNumPyPandasFlaskTokenizerLLMDockerOpen API

Cloud Technology

AzureAzure FunctionsAzure storageGitHubGitHub actions

Industry

Open Source

Overview

In the age of artificial intelligence, we proudly introduce "IsROBERTa" — our ground-breaking achievement as the first Icelandic large transformer language model. Developed in 2020, IsROBERTa was designed to understand and generate Icelandic text, pushing the boundaries of language processing in AI for Icelandic.

Challenges

Creating a large transformer language model, particularly for a language as unique and complex as Icelandic, is no small feat. The field of training large language models in 2020 was still in its infancy. Some challenges we faced were:

  • Data Scarcity
  • Language Complexity
  • Computational Resources
  • Lack of established libraries

About the model

IsROBERTa was trained on the OSCAR corpus using a Masked Language Model (MLM) downstream task, boasting the following hyperparameters:

  • Per-device training batch size: 48
  • Number of epochs: 1
  • Vocabulary size: 52,000
  • Maximum position embeddings: 514
  • Number of attention heads: 12
  • Number of hidden layers: 6
  • Type vocabulary size: 1
  • Learning rate: 0.00005

For detailed insights into the model, visit our GitHub repository.

Your model

We believe that the journey towards breakthrough AI solutions is best embarked on together. When you choose to work with us, you're not just selecting a service provider, you're choosing a partner dedicated to bringing your AI vision to life.