update readme
Browse files- README.md +6 -10
- vllm_plugin_meralion/README.md +1 -1
README.md
CHANGED
@@ -406,17 +406,17 @@ Refer to the [AudioBench paper](https://arxiv.org/abs/2406.16020) for more detai
|
|
406 |
</table>
|
407 |
</div>
|
408 |
|
409 |
-
##
|
410 |
-
|
411 |
-
Here we provide a code snippet illustrating the process of loading both the processor and model, alongside detailed instructions on executing the MERaLiON-AudioLLM model for content generation.
|
412 |
|
413 |
> [!WARNING]
|
414 |
> **Out of Scope use**: This model is not intended for use in tool calling, math, and coding tasks.
|
415 |
|
|
|
|
|
|
|
416 |
|
417 |
-
### CPU Inference
|
418 |
|
419 |
-
|
420 |
|
421 |
```python
|
422 |
import librosa
|
@@ -459,7 +459,7 @@ generated_ids = outputs[:, inputs['input_ids'].size(1):]
|
|
459 |
response = processor.batch_decode(generated_ids, skip_special_tokens=True)
|
460 |
```
|
461 |
|
462 |
-
### GPU Inference
|
463 |
|
464 |
```python
|
465 |
import torch
|
@@ -513,10 +513,6 @@ generated_ids = outputs[:, inputs['input_ids'].size(1):]
|
|
513 |
response = processor.batch_decode(generated_ids, skip_special_tokens=True)
|
514 |
```
|
515 |
|
516 |
-
### vLLM Inference
|
517 |
-
|
518 |
-
We support hosting the model using vLLM framework. Refer to the guide [here](vllm_plugin_meralion/README.md).
|
519 |
-
|
520 |
## Disclaimer
|
521 |
|
522 |
The current MERaLiON-AudioLLM has not been specifically aligned for safety and may generate content that is inappropriate, offensive, or harmful. Developers and users are responsible for performing their own safety fine-tuning and implementing necessary security measures. The authors shall not be held liable for any claims, damages, or other liabilities arising from the use of the released models, weights, or code.
|
|
|
406 |
</table>
|
407 |
</div>
|
408 |
|
409 |
+
## Inference
|
|
|
|
|
410 |
|
411 |
> [!WARNING]
|
412 |
> **Out of Scope use**: This model is not intended for use in tool calling, math, and coding tasks.
|
413 |
|
414 |
+
### vLLM Inference
|
415 |
+
|
416 |
+
We support hosting the model using vLLM framework. Refer to the guide [here](vllm_plugin_meralion/README.md).
|
417 |
|
|
|
418 |
|
419 |
+
### Huggingface CPU Inference
|
420 |
|
421 |
```python
|
422 |
import librosa
|
|
|
459 |
response = processor.batch_decode(generated_ids, skip_special_tokens=True)
|
460 |
```
|
461 |
|
462 |
+
### Huggingface GPU Inference
|
463 |
|
464 |
```python
|
465 |
import torch
|
|
|
513 |
response = processor.batch_decode(generated_ids, skip_special_tokens=True)
|
514 |
```
|
515 |
|
|
|
|
|
|
|
|
|
516 |
## Disclaimer
|
517 |
|
518 |
The current MERaLiON-AudioLLM has not been specifically aligned for safety and may generate content that is inappropriate, offensive, or harmful. Developers and users are responsible for performing their own safety fine-tuning and implementing necessary security measures. The authors shall not be held liable for any claims, damages, or other liabilities arising from the use of the released models, weights, or code.
|
vllm_plugin_meralion/README.md
CHANGED
@@ -13,7 +13,7 @@ pip install transformers==4.46.3
|
|
13 |
```
|
14 |
|
15 |
As the [vLLM documentation](https://docs.vllm.ai/en/stable/models/adding_model.html#out-of-tree-model-integration) recommends,
|
16 |
-
we provide a way to register our model via [vLLM plugins](https://docs.vllm.ai/en/stable/design/plugin_system.html#plugin-system).
|
17 |
|
18 |
|
19 |
```bash
|
|
|
13 |
```
|
14 |
|
15 |
As the [vLLM documentation](https://docs.vllm.ai/en/stable/models/adding_model.html#out-of-tree-model-integration) recommends,
|
16 |
+
we provide a way to register our model via [vLLM plugins](https://docs.vllm.ai/en/stable/design/plugin_system.html#plugin-system). Simply install by running the following command.
|
17 |
|
18 |
|
19 |
```bash
|