mox's picture
Update README.md
2c38fa7
metadata
language:
  - de
tags:
  - Text Classification
  - sentiment
  - Simpletransformers
  - deepset/gbert-base

This gBert-base model was finetuned on a sentiment prediction task with tweets from German politician during the German Federal Election in 2021.

Model Description:

This model was trained on ~30.000 annotated tweets in German language on its sentiment. It can predict tweets as negative, positive or neutral. It achieved an accuracy of 93% on the specific dataset.

Model Implementation

You can implement this model for example with Simpletransformers. First you have to unpack the file.

def unpack_model(model_name=''):
  tar = tarfile.open(f"{model_name}.tar.gz", "r:gz")
  tar.extractall()
  tar.close()
  

The hyperparameter were defined as follows: train_args ={"reprocess_input_data": True, "fp16":False, "num_train_epochs": 4, "overwrite_output_dir":True, "train_batch_size": 32, "eval_batch_size": 32}

Now create the model: unpack_model(YOUR_DOWNLOADED_FILE_HERE)

model = ClassificationModel(
    "bert", "content/outputs/",
    num_labels= 3,
    args=train_args
)

In this case for the output:

  • 0 = positive
  • 1 = negative
  • 2 = neutral

Example for a positive prediction:

model.predict(["Das ist gut! Wir danken dir."])
([0], array([[ 2.06561327, -3.57908797,  1.5340755 ]]))

Example for a negative prediction:

model.predict(["Ich hasse dich!"])
([1], array([[-3.50486898,  4.29590368, -0.9000684 ]]))

Example for a neutral prediction:

model.predict(["Heute ist Sonntag."])
([2], array([[-2.94458342, -2.91875601,  4.94414234]]))

This model was created by Maximilian Weissenbacher for a project at the University of Regensburg.