Spaces:
Running
Running
| ## DALL-E Mini - Generate image from text | |
| ## TODO | |
| * experiment with flax/jax and setup of the TPU instance that we should get shortly | |
| * work on dataset loading - [see suggested datasets](https://discuss.huggingface.co/t/dall-e-mini-version/7324/4) | |
| * Optionally create the OpenAI YFCC100M subset (see [this post](https://discuss.huggingface.co/t/dall-e-mini-version/7324/30?u=boris)) | |
| * work on text/image encoding | |
| * concatenate inputs (not sure if we need fixed length for text or use a special token separating text & image) | |
| * adapt training script | |
| * create inference function | |
| * integrate CLIP for better results (only if we have the time) | |
| * work on a demo (streamlit or colab or maybe just HF widget) | |
| * document (set up repo on model hub per instructions, start on README writeup…) | |
| * help with coordinating activities & progress | |
| ## Dependencies Installation | |
| You should create a new python virtual environment and install the project dependencies inside the virtual env: `pip install -r requirements.txt` | |
| If you use `conda`, you can create the virtual env and install everything using: `conda env update -f environments.yaml` | |