add model
Browse files- README.md +57 -0
 - config.json +39 -0
 - tf_model.h5 +3 -0
 
    	
        README.md
    ADDED
    
    | 
         @@ -0,0 +1,57 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            ---
         
     | 
| 2 | 
         
            +
            license: apache-2.0
         
     | 
| 3 | 
         
            +
            tags:
         
     | 
| 4 | 
         
            +
            - generated_from_keras_callback
         
     | 
| 5 | 
         
            +
            model-index:
         
     | 
| 6 | 
         
            +
            - name: relevance-model
         
     | 
| 7 | 
         
            +
              results: []
         
     | 
| 8 | 
         
            +
            ---
         
     | 
| 9 | 
         
            +
             
     | 
| 10 | 
         
            +
            <!-- This model card has been generated automatically according to the information Keras had access to. You should
         
     | 
| 11 | 
         
            +
            probably proofread and complete it, then remove this comment. -->
         
     | 
| 12 | 
         
            +
             
     | 
| 13 | 
         
            +
            # relevance-model
         
     | 
| 14 | 
         
            +
             
     | 
| 15 | 
         
            +
            This model is a fine-tuned version of [hfl/chinese-roberta-wwm-ext](https://huggingface.co/hfl/chinese-roberta-wwm-ext) on an unknown dataset.
         
     | 
| 16 | 
         
            +
            It achieves the following results on the evaluation set:
         
     | 
| 17 | 
         
            +
            - Train Loss: 0.3134
         
     | 
| 18 | 
         
            +
            - Train Binary Accuracy: 0.8773
         
     | 
| 19 | 
         
            +
            - Validation Loss: 0.3633
         
     | 
| 20 | 
         
            +
            - Validation Binary Accuracy: 0.8541
         
     | 
| 21 | 
         
            +
            - Epoch: 2
         
     | 
| 22 | 
         
            +
             
     | 
| 23 | 
         
            +
            ## Model description
         
     | 
| 24 | 
         
            +
             
     | 
| 25 | 
         
            +
            More information needed
         
     | 
| 26 | 
         
            +
             
     | 
| 27 | 
         
            +
            ## Intended uses & limitations
         
     | 
| 28 | 
         
            +
             
     | 
| 29 | 
         
            +
            More information needed
         
     | 
| 30 | 
         
            +
             
     | 
| 31 | 
         
            +
            ## Training and evaluation data
         
     | 
| 32 | 
         
            +
             
     | 
| 33 | 
         
            +
            More information needed
         
     | 
| 34 | 
         
            +
             
     | 
| 35 | 
         
            +
            ## Training procedure
         
     | 
| 36 | 
         
            +
             
     | 
| 37 | 
         
            +
            ### Training hyperparameters
         
     | 
| 38 | 
         
            +
             
     | 
| 39 | 
         
            +
            The following hyperparameters were used during training:
         
     | 
| 40 | 
         
            +
            - optimizer: {'name': 'Adam', 'learning_rate': 5e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}
         
     | 
| 41 | 
         
            +
            - training_precision: float32
         
     | 
| 42 | 
         
            +
             
     | 
| 43 | 
         
            +
            ### Training results
         
     | 
| 44 | 
         
            +
             
     | 
| 45 | 
         
            +
            | Train Loss | Train Binary Accuracy | Validation Loss | Validation Binary Accuracy | Epoch |
         
     | 
| 46 | 
         
            +
            |:----------:|:---------------------:|:---------------:|:--------------------------:|:-----:|
         
     | 
| 47 | 
         
            +
            | 0.3980     | 0.8289                | 0.3739          | 0.8541                     | 0     |
         
     | 
| 48 | 
         
            +
            | 0.3446     | 0.8606                | 0.3614          | 0.8505                     | 1     |
         
     | 
| 49 | 
         
            +
            | 0.3134     | 0.8773                | 0.3633          | 0.8541                     | 2     |
         
     | 
| 50 | 
         
            +
             
     | 
| 51 | 
         
            +
             
     | 
| 52 | 
         
            +
            ### Framework versions
         
     | 
| 53 | 
         
            +
             
     | 
| 54 | 
         
            +
            - Transformers 4.16.0
         
     | 
| 55 | 
         
            +
            - TensorFlow 2.7.0
         
     | 
| 56 | 
         
            +
            - Datasets 1.18.1
         
     | 
| 57 | 
         
            +
            - Tokenizers 0.11.0
         
     | 
    	
        config.json
    ADDED
    
    | 
         @@ -0,0 +1,39 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "_name_or_path": "hfl/chinese-roberta-wwm-ext",
         
     | 
| 3 | 
         
            +
              "architectures": [
         
     | 
| 4 | 
         
            +
                "BertForSequenceClassification"
         
     | 
| 5 | 
         
            +
              ],
         
     | 
| 6 | 
         
            +
              "attention_probs_dropout_prob": 0.1,
         
     | 
| 7 | 
         
            +
              "bos_token_id": 0,
         
     | 
| 8 | 
         
            +
              "classifier_dropout": null,
         
     | 
| 9 | 
         
            +
              "directionality": "bidi",
         
     | 
| 10 | 
         
            +
              "eos_token_id": 2,
         
     | 
| 11 | 
         
            +
              "hidden_act": "gelu",
         
     | 
| 12 | 
         
            +
              "hidden_dropout_prob": 0.1,
         
     | 
| 13 | 
         
            +
              "hidden_size": 768,
         
     | 
| 14 | 
         
            +
              "id2label": {
         
     | 
| 15 | 
         
            +
                "0": "LABEL_0"
         
     | 
| 16 | 
         
            +
              },
         
     | 
| 17 | 
         
            +
              "initializer_range": 0.02,
         
     | 
| 18 | 
         
            +
              "intermediate_size": 3072,
         
     | 
| 19 | 
         
            +
              "label2id": {
         
     | 
| 20 | 
         
            +
                "LABEL_0": 0
         
     | 
| 21 | 
         
            +
              },
         
     | 
| 22 | 
         
            +
              "layer_norm_eps": 1e-12,
         
     | 
| 23 | 
         
            +
              "max_position_embeddings": 512,
         
     | 
| 24 | 
         
            +
              "model_type": "bert",
         
     | 
| 25 | 
         
            +
              "num_attention_heads": 12,
         
     | 
| 26 | 
         
            +
              "num_hidden_layers": 12,
         
     | 
| 27 | 
         
            +
              "output_past": true,
         
     | 
| 28 | 
         
            +
              "pad_token_id": 1,
         
     | 
| 29 | 
         
            +
              "pooler_fc_size": 768,
         
     | 
| 30 | 
         
            +
              "pooler_num_attention_heads": 12,
         
     | 
| 31 | 
         
            +
              "pooler_num_fc_layers": 3,
         
     | 
| 32 | 
         
            +
              "pooler_size_per_head": 128,
         
     | 
| 33 | 
         
            +
              "pooler_type": "first_token_transform",
         
     | 
| 34 | 
         
            +
              "position_embedding_type": "absolute",
         
     | 
| 35 | 
         
            +
              "transformers_version": "4.16.0",
         
     | 
| 36 | 
         
            +
              "type_vocab_size": 2,
         
     | 
| 37 | 
         
            +
              "use_cache": true,
         
     | 
| 38 | 
         
            +
              "vocab_size": 21128
         
     | 
| 39 | 
         
            +
            }
         
     | 
    	
        tf_model.h5
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:cbab7d1305de4cb3d153551654e65b24899cb2ccbef62456cefb3789d4d28fce
         
     | 
| 3 | 
         
            +
            size 409361748
         
     |