YingxuHe commited on
Commit
b9ddb88
·
1 Parent(s): d5c5e9f

update readme

Browse files
Files changed (2) hide show
  1. README.md +6 -10
  2. vllm_plugin_meralion/README.md +1 -1
README.md CHANGED
@@ -406,17 +406,17 @@ Refer to the [AudioBench paper](https://arxiv.org/abs/2406.16020) for more detai
406
  </table>
407
  </div>
408
 
409
- ## Uses
410
-
411
- Here we provide a code snippet illustrating the process of loading both the processor and model, alongside detailed instructions on executing the MERaLiON-AudioLLM model for content generation.
412
 
413
  > [!WARNING]
414
  > **Out of Scope use**: This model is not intended for use in tool calling, math, and coding tasks.
415
 
 
 
 
416
 
417
- ### CPU Inference
418
 
419
- MERaLiON-AudioLLM also supports batch inference.
420
 
421
  ```python
422
  import librosa
@@ -459,7 +459,7 @@ generated_ids = outputs[:, inputs['input_ids'].size(1):]
459
  response = processor.batch_decode(generated_ids, skip_special_tokens=True)
460
  ```
461
 
462
- ### GPU Inference
463
 
464
  ```python
465
  import torch
@@ -513,10 +513,6 @@ generated_ids = outputs[:, inputs['input_ids'].size(1):]
513
  response = processor.batch_decode(generated_ids, skip_special_tokens=True)
514
  ```
515
 
516
- ### vLLM Inference
517
-
518
- We support hosting the model using vLLM framework. Refer to the guide [here](vllm_plugin_meralion/README.md).
519
-
520
  ## Disclaimer
521
 
522
  The current MERaLiON-AudioLLM has not been specifically aligned for safety and may generate content that is inappropriate, offensive, or harmful. Developers and users are responsible for performing their own safety fine-tuning and implementing necessary security measures. The authors shall not be held liable for any claims, damages, or other liabilities arising from the use of the released models, weights, or code.
 
406
  </table>
407
  </div>
408
 
409
+ ## Inference
 
 
410
 
411
  > [!WARNING]
412
  > **Out of Scope use**: This model is not intended for use in tool calling, math, and coding tasks.
413
 
414
+ ### vLLM Inference
415
+
416
+ We support hosting the model using vLLM framework. Refer to the guide [here](vllm_plugin_meralion/README.md).
417
 
 
418
 
419
+ ### Huggingface CPU Inference
420
 
421
  ```python
422
  import librosa
 
459
  response = processor.batch_decode(generated_ids, skip_special_tokens=True)
460
  ```
461
 
462
+ ### Huggingface GPU Inference
463
 
464
  ```python
465
  import torch
 
513
  response = processor.batch_decode(generated_ids, skip_special_tokens=True)
514
  ```
515
 
 
 
 
 
516
  ## Disclaimer
517
 
518
  The current MERaLiON-AudioLLM has not been specifically aligned for safety and may generate content that is inappropriate, offensive, or harmful. Developers and users are responsible for performing their own safety fine-tuning and implementing necessary security measures. The authors shall not be held liable for any claims, damages, or other liabilities arising from the use of the released models, weights, or code.
vllm_plugin_meralion/README.md CHANGED
@@ -13,7 +13,7 @@ pip install transformers==4.46.3
13
  ```
14
 
15
  As the [vLLM documentation](https://docs.vllm.ai/en/stable/models/adding_model.html#out-of-tree-model-integration) recommends,
16
- we provide a way to register our model via [vLLM plugins](https://docs.vllm.ai/en/stable/design/plugin_system.html#plugin-system).
17
 
18
 
19
  ```bash
 
13
  ```
14
 
15
  As the [vLLM documentation](https://docs.vllm.ai/en/stable/models/adding_model.html#out-of-tree-model-integration) recommends,
16
+ we provide a way to register our model via [vLLM plugins](https://docs.vllm.ai/en/stable/design/plugin_system.html#plugin-system). Simply install by running the following command.
17
 
18
 
19
  ```bash