kaleido-base / README.md
tsor13's picture
Update README.md
2bc2ea2
---
license: apache-2.0
tags:
- generated_from_trainer
model-index:
- name: flan-t5-base-a100
results: []
widget:
- text: "[Generate]:\tSituation: Do a pushup"
example_title: Generate
- text: "[Relevance]:\tSituation: Do a pushup\tValue: Health"
example_title: Relevance
- text: "[Valence]:\tSituation: Do a pushup\tValue: Health"
example_title: Valence
- text: "[Explanation]:\tSituation: Do a pushup\tValue: Health"
example_title: Explain
datasets:
- allenai/ValuePrism
extra_gated_prompt: >-
Access to this model is automatically granted upon accepting the [**AI2
ImpACT License - Medium Risk Artifacts (“MR
Agreement”)**](https://allenai.org/licenses/impact-mr) and completing all
fields below.
extra_gated_fields:
Your full name: text
Organization or entity you are affiliated with: text
State or country you are located in: text
Contact email: text
Please describe your intended use of the medium risk artifact(s): text
I UNDERSTAND that the model is intended for research purposes and not for real-world use-cases: checkbox
I AGREE to the terms and conditions of the MR Agreement above: checkbox
I AGREE to AI2’s use of my information for legal notices and administrative matters: checkbox
I CERTIFY that the information I have provided is true and accurate: checkbox
---
# Model Card for Kaleido
## Model Description
Kaleido is a multi-task seq2seq model designed to _generate_, _explain_, and output the _relevance_ and _valence_ of contextualized values, rights, and duties, distilled from GPT-4 generated data.
- **Model type:** Language model
- **Language(s) (NLP):** en
- **License:** [AI2
ImpACT License - Medium Risk Artifacts (“MR
Agreement”)](https://allenai.org/licenses/impact-mr)
- **Parent Model:** [google/flan-t5-base](https://huggingface.co/google/flan-t5-base)
- **GitHub Repo:** https://github.com/tsor13/kaleido
- **Paper:** https://arxiv.org/abs/2309.00779
- **Demo:** https://kaleido.allen.ai
- **All model sizes:** [small](https://huggingface.co/tsor13/kaleido-small), [base](https://huggingface.co/tsor13/kaleido-base), [large](https://huggingface.co/tsor13/kaleido-large), [xl](https://huggingface.co/tsor13/kaleido-xl), [xxl](https://huggingface.co/tsor13/kaleido-xxl)
# Uses
It is intended to be used for research purposes to a) understand how well large language models can approximate pluralistic human values and b) to make an open, transparent attempt to increase the capabilities of LLMs to model human values.
## Out-of-Scope Use
The model is not intended to be used for advice, human-facing applications, or other purposes.
# Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
Additionally, certain groups may be represented better in the model's outputs than others, and the fact that the data is entirely in English and generated by predominantly by English speakers/LLMs trained on English, the model's outputs likely fit perspectives from English-speaking countries better.
The relevance score _should not_ be interpreted as an importance score, but, due to the composition of the training data, corresponds more closely with "is this value likely to have been generated for this situation by GPT-4?"
## Recommendations
We recommend that this model not be used for any high-impact or human-facing purposes as its biases and limitations need to be further explored.
We intend this to be a research artifact to advance AI's ability to model and interact with pluralistic human values, rights, and duties.
# Training Details
## Training Data
The model is trained on the `mixture` training split of [ValuePrism](https://huggingface.co/datasets/tsor13/ValuePrism).
## Training Procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 32
- eval_batch_size: 8
- seed: 42
- distributed_type: multi-GPU
- num_devices: 2
- total_train_batch_size: 64
- total_eval_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 4.0
### Framework versions
- Transformers 4.22.0.dev0
- Pytorch 1.12.1+cu113
- Datasets 2.4.0
- Tokenizers 0.12.1
## Testing Data, Factors & Metrics
### Testing Data
The model is tested on the four subtasks in [ValuePrism](https://huggingface.co/datasets/tsor13/ValuePrism).
### Metrics
Accuracy is used for relevance and valence, as they are classification tasks, and perplexity is used for generation and explanation, as they are free text generation tasks.
## Results
| **Model** | **Relevance Acc ↑** | **Valence Acc ↑** | **Generative Perp ↓** | **Explanation Perp ↓** |
|-------------------------|------------------|-------------------|-----------------|------------------|
| `kaleido-xxl` (11B) | 89.1 | 81.9 | 2.22 | 2.99 |
| `kaleido-xl` (3B) | 88.4 | 80.8 | 2.23 | 3.14 |
| `kaleido-large` (770M) | 87.2 | 79.2 | 2.34 | 3.52 |
| `kaleido-base` (220M) | 83.5 | 74.5 | 2.53 | 4.23 |
| `kaleido-small` (60M) | 66.0 | 59.7 | 2.86 | 5.70 |
## Model Architecture and Objective
Kaleido is an encoder-decoder T5-based model trained using negative log-likelihood.
# Citation
**BibTeX:**
```
@misc{sorensen2023value,
title={Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties},
author={Taylor Sorensen and Liwei Jiang and Jena Hwang and Sydney Levine and Valentina Pyatkin and Peter West and Nouha Dziri and Ximing Lu and Kavel Rao and Chandra Bhagavatula and Maarten Sap and John Tasioulas and Yejin Choi},
year={2023},
eprint={2309.00779},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
# Model Card Contact
Contact Taylor Sorensen (`[email protected]`) for any questions about this model.
# How to Get Started with the Model
Use the code below to get started with the model.
**Load the model:**
```python
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
model_name = 'allenai/kaleido-base'
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSeq2SeqLM.from_pretrained(model_name)
```
Each task ('generate', 'relevance', 'valence', 'explanation') has its own template that can be accessed from the model config.
```python
task = 'generate' # can be 'generate', 'relevance', 'valence', or 'explanation'
model.config.task_specific_params[task]['template']
```
Output:
```
'[Generate]:\tAction: ACTION'
```
The generate template requires `ACTION`, while the other three templates require `ACTION`, `VRD` (`'Value', 'Right', or 'Duty'`) and `TEXT`.
Replace the arguments with the text and generate with the model.
Generate example:
```python
action = 'Go to the gym'
input_text = model.config.task_specific_params['generate']['template'].replace('ACTION', action)
# tokenize
input_ids = tokenizer.encode(input_text, return_tensors='pt')
# generate
output_ids = model.generate(input_ids, max_length=64)
# decode
output_text = tokenizer.decode(output_ids[0], skip_special_tokens=True)
output_text
```
Output:
```
'Value: Personal growth'
```
Valence example:
```python
action = 'Go to the gym'
vrd = 'Value'
text = 'Health'
input_text = model.config.task_specific_params['valence']['template']
replacements = {
'ACTION': action,
'VRD': vrd,
'TEXT': text
}
for key, value in replacements.items():
input_text = input_text.replace(key, value)
# tokenize
input_ids = tokenizer.encode(input_text, return_tensors='pt')
# generate
output_ids = model.generate(input_ids, max_length=64)
# decode
output_text = tokenizer.decode(output_ids[0], skip_special_tokens=True)
output_text
```
Output:
```
'Supports'
```
Alternatively, one could use the `KaleidoSys` class in the companion [git repo](https://github.com/tsor13/kaleido/tree/main) which automates templatization and such.