--- language: - de tags: - Text Classification - sentiment - Simpletransformers - deepset/gbert-base --- This gBert-base model was finetuned on a sentiment prediction task with tweets from German politician during the German Federal Election in 2021. ## Model Description: This model was trained on ~30.000 annotated tweets in German language on its sentiment. It can predict tweets as negative, positive or neutral. It achieved an accuracy of 93% on the specific dataset. ## Model Implementation You can implement this model for example with Simpletransformers. First you have to unpack the file. def unpack_model(model_name=''): tar = tarfile.open(f"{model_name}.tar.gz", "r:gz") tar.extractall() tar.close() The hyperparameter were defined as follows: train_args ={"reprocess_input_data": True, "fp16":False, "num_train_epochs": 4, "overwrite_output_dir":True, "train_batch_size": 32, "eval_batch_size": 32} Now create the model: unpack_model(YOUR_DOWNLOADED_FILE_HERE) model = ClassificationModel( "bert", "content/outputs/", num_labels= 3, args=train_args ) In this case for the output: - 0 = positive - 1 = negative - 2 = neutral Example for a positive prediction: model.predict(["Das ist gut! Wir danken dir."]) ([0], array([[ 2.06561327, -3.57908797, 1.5340755 ]])) Example for a negative prediction: model.predict(["Ich hasse dich!"]) ([1], array([[-3.50486898, 4.29590368, -0.9000684 ]])) Example for a neutral prediction: model.predict(["Heute ist Sonntag."]) ([2], array([[-2.94458342, -2.91875601, 4.94414234]])) This model was created by Maximilian Weissenbacher for a project at the University of Regensburg.