YingxuHe commited on
Commit
5cd6cd7
·
verified ·
1 Parent(s): e2042c3

Upload config

Browse files
Files changed (2) hide show
  1. README.md +18 -7
  2. config.json +1 -8
README.md CHANGED
@@ -1,4 +1,8 @@
1
  ---
 
 
 
 
2
  library_name: transformers
3
  license: other
4
  license_name: meralion-public-license
@@ -13,23 +17,30 @@ tags:
13
  - chat
14
  - audio
15
  - safetensors
16
- datasets:
17
- - MERaLiON/MNSC
18
- base_model:
19
- - openai/whisper-large-v2
20
  widget:
21
  - example_title: Sentence-level ASR
22
  src: librispeech_clean.wav
23
  output:
24
- text: "USER Instruction: Please transcribe this speech.\n MODEL Output: <Speaker1> When they were outside, Ung simply latched the door and started up the path."
 
25
  - example_title: Dialogue-level ASR
26
  src: IMDA_conversation.wav
27
  output:
28
- text: "USER Instruction: Please turn this speech into written format.\n MODEL Output: <Speaker1> Okay, (um) in that case, (uh) how do I apply? <Speaker2> Alright, you can just (um) apply it online and then (um) we'll need some documents from you. (um) let's say the bank statement with your address and your name, and also the (um) Nsf card that you have to enjoy the promotion. <Speaker1> (mmhmm) (mmhmm) (mmhmm) [ah] I see the green one, right?"
 
 
 
 
 
29
  - example_title: Spoken Dialogue Summarization
30
  src: IMDA_conversation.wav
31
  output:
32
- text: "USER Instruction: Please briefly summarize this conversation.\n MODEL Output: Speaker1 and Speaker2 discussed the application process for a promotion. Speaker2 explained that the application can be done online and provided a list of required documents, including a bank statement with the applicant's address and name, and an NSF card. Speaker1 acknowledged the information and confirmed the details of the required documents."
 
 
 
 
 
33
  ---
34
 
35
  # MERaLiON
 
1
  ---
2
+ base_model:
3
+ - openai/whisper-large-v2
4
+ datasets:
5
+ - MERaLiON/MNSC
6
  library_name: transformers
7
  license: other
8
  license_name: meralion-public-license
 
17
  - chat
18
  - audio
19
  - safetensors
 
 
 
 
20
  widget:
21
  - example_title: Sentence-level ASR
22
  src: librispeech_clean.wav
23
  output:
24
+ text: "USER Instruction: Please transcribe this speech.\n MODEL Output: <Speaker1>\
25
+ \ When they were outside, Ung simply latched the door and started up the path."
26
  - example_title: Dialogue-level ASR
27
  src: IMDA_conversation.wav
28
  output:
29
+ text: "USER Instruction: Please turn this speech into written format.\n MODEL\
30
+ \ Output: <Speaker1> Okay, (um) in that case, (uh) how do I apply? <Speaker2>\
31
+ \ Alright, you can just (um) apply it online and then (um) we'll need some documents\
32
+ \ from you. (um) let's say the bank statement with your address and your name,\
33
+ \ and also the (um) Nsf card that you have to enjoy the promotion. <Speaker1>\
34
+ \ (mmhmm) (mmhmm) (mmhmm) [ah] I see the green one, right?"
35
  - example_title: Spoken Dialogue Summarization
36
  src: IMDA_conversation.wav
37
  output:
38
+ text: "USER Instruction: Please briefly summarize this conversation.\n MODEL Output:\
39
+ \ Speaker1 and Speaker2 discussed the application process for a promotion. Speaker2\
40
+ \ explained that the application can be done online and provided a list of required\
41
+ \ documents, including a bank statement with the applicant's address and name,\
42
+ \ and an NSF card. Speaker1 acknowledged the information and confirmed the details\
43
+ \ of the required documents."
44
  ---
45
 
46
  # MERaLiON
config.json CHANGED
@@ -1,10 +1,6 @@
1
  {
2
- "architectures": [
3
- "MERaLiONForConditionalGeneration"
4
- ],
5
  "auto_map": {
6
- "AutoConfig": "configuration_meralion.MERaLiONConfig",
7
- "AutoModelForSpeechSeq2Seq": "modeling_meralion.MERaLiONForConditionalGeneration"
8
  },
9
  "head_dim": 256,
10
  "hidden_size": 3584,
@@ -15,7 +11,6 @@
15
  "num_key_value_heads": 8,
16
  "sliding_window": 4096,
17
  "speech_config": {
18
- "_attn_implementation_autoset": true,
19
  "_name_or_path": "openai/whisper-large-v2",
20
  "apply_spec_augment": true,
21
  "architectures": [
@@ -150,7 +145,6 @@
150
  "speech_mlp_scale_factor": 15,
151
  "speech_token_index": 255999,
152
  "text_config": {
153
- "_attn_implementation_autoset": true,
154
  "_name_or_path": "aisingapore/gemma2-9b-cpt-sea-lionv3-instruct",
155
  "architectures": [
156
  "Gemma2ForCausalLM"
@@ -166,6 +160,5 @@
166
  "sliding_window_size": 4096,
167
  "torch_dtype": "bfloat16"
168
  },
169
- "torch_dtype": "bfloat16",
170
  "transformers_version": "4.46.3"
171
  }
 
1
  {
 
 
 
2
  "auto_map": {
3
+ "AutoConfig": "configuration_meralion.MERaLiONConfig"
 
4
  },
5
  "head_dim": 256,
6
  "hidden_size": 3584,
 
11
  "num_key_value_heads": 8,
12
  "sliding_window": 4096,
13
  "speech_config": {
 
14
  "_name_or_path": "openai/whisper-large-v2",
15
  "apply_spec_augment": true,
16
  "architectures": [
 
145
  "speech_mlp_scale_factor": 15,
146
  "speech_token_index": 255999,
147
  "text_config": {
 
148
  "_name_or_path": "aisingapore/gemma2-9b-cpt-sea-lionv3-instruct",
149
  "architectures": [
150
  "Gemma2ForCausalLM"
 
160
  "sliding_window_size": 4096,
161
  "torch_dtype": "bfloat16"
162
  },
 
163
  "transformers_version": "4.46.3"
164
  }