google-bert / bert-base-german-cased

huggingface.co
Total runs: 479.2K
24-hour runs: 2.5K
7-day runs: 8.0K
30-day runs: 3.0K
Model's Last Updated: February 19 2024
fill-mask

Introduction of bert-base-german-cased

Model Details of bert-base-german-cased

German BERT

bert_image

Overview

Language model: bert-base-cased
Language: German
Training data: Wiki, OpenLegalData, News (~ 12GB)
Eval data: Conll03 (NER), GermEval14 (NER), GermEval18 (Classification), GNAD (Classification)
Infrastructure : 1x TPU v2
Published : Jun 14th, 2019

Update April 3rd, 2020 : we updated the vocabulary file on deepset's s3 to conform with the default tokenization of punctuation tokens. For details see the related FARM issue . If you want to use the old vocab we have also uploaded a "deepset/bert-base-german-cased-oldvocab" model.

Details
  • We trained using Google's Tensorflow code on a single cloud TPU v2 with standard settings.
  • We trained 810k steps with a batch size of 1024 for sequence length 128 and 30k steps with sequence length 512. Training took about 9 days.
  • As training data we used the latest German Wikipedia dump (6GB of raw txt files), the OpenLegalData dump (2.4 GB) and news articles (3.6 GB).
  • We cleaned the data dumps with tailored scripts and segmented sentences with spacy v2.1. To create tensorflow records we used the recommended sentencepiece library for creating the word piece vocabulary and tensorflow scripts to convert the text to data usable by BERT.

See https://deepset.ai/german-bert for more details

Hyperparameters
batch_size = 1024
n_steps = 810_000
max_seq_len = 128 (and 512 later)
learning_rate = 1e-4
lr_schedule = LinearWarmup
num_warmup_steps = 10_000
Performance

During training we monitored the loss and evaluated different model checkpoints on the following German datasets:

  • germEval18Fine: Macro f1 score for multiclass sentiment classification
  • germEval18coarse: Macro f1 score for binary sentiment classification
  • germEval14: Seq f1 score for NER (file names deuutf.*)
  • CONLL03: Seq f1 score for NER
  • 10kGNAD: Accuracy for document classification

Even without thorough hyperparameter tuning, we observed quite stable learning especially for our German model. Multiple restarts with different seeds produced quite similar results.

performancetable

We further evaluated different points during the 9 days of pre-training and were astonished how fast the model converges to the maximally reachable performance. We ran all 5 downstream tasks on 7 different model checkpoints - taken at 0 up to 840k training steps (x-axis in figure below). Most checkpoints are taken from early training where we expected most performance changes. Surprisingly, even a randomly initialized BERT can be trained only on labeled downstream datasets and reach good performance (blue line, GermEval 2018 Coarse task, 795 kB trainset size).

checkpointseval

Authors
  • Branden Chan: branden.chan [at] deepset.ai
  • Timo Möller: timo.moeller [at] deepset.ai
  • Malte Pietsch: malte.pietsch [at] deepset.ai
  • Tanay Soni: tanay.soni [at] deepset.ai
About us

deepset logo

We bring NLP to the industry via open source!
Our focus: Industry specific language models & large scale QA systems.

Some of our work:

Get in touch: Twitter | LinkedIn | Website

Runs of google-bert bert-base-german-cased on huggingface.co

479.2K
Total runs
2.5K
24-hour runs
1.1K
3-day runs
8.0K
7-day runs
3.0K
30-day runs

More Information About bert-base-german-cased huggingface.co Model

More bert-base-german-cased license Visit here:

https://choosealicense.com/licenses/mit

bert-base-german-cased huggingface.co

bert-base-german-cased huggingface.co is an AI model on huggingface.co that provides bert-base-german-cased's model effect (), which can be used instantly with this google-bert bert-base-german-cased model. huggingface.co supports a free trial of the bert-base-german-cased model, and also provides paid use of the bert-base-german-cased. Support call bert-base-german-cased model through api, including Node.js, Python, http.

bert-base-german-cased huggingface.co Url

https://huggingface.co/google-bert/bert-base-german-cased

google-bert bert-base-german-cased online free

bert-base-german-cased huggingface.co is an online trial and call api platform, which integrates bert-base-german-cased's modeling effects, including api services, and provides a free online trial of bert-base-german-cased, you can try bert-base-german-cased online for free by clicking the link below.

google-bert bert-base-german-cased online free url in huggingface.co:

https://huggingface.co/google-bert/bert-base-german-cased

bert-base-german-cased install

bert-base-german-cased is an open source model from GitHub that offers a free installation service, and any user can find bert-base-german-cased on GitHub to install. At the same time, huggingface.co provides the effect of bert-base-german-cased install, users can directly use bert-base-german-cased installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

bert-base-german-cased install url in huggingface.co:

https://huggingface.co/google-bert/bert-base-german-cased

Url of bert-base-german-cased

bert-base-german-cased huggingface.co Url

Provider of bert-base-german-cased huggingface.co

google-bert
ORGANIZATIONS

Other API from google-bert