--- license: mit widget: - text: I am having itching, skin rash, and nodal skin eruptions example_title: Fungal infection example - text: I feel like vomiting, breathlessness, and sweating example_title: Heart Attack example - text: I am feeling fatigue, weight loss, restlessness and also lethargy. example_title: Diabetes example --- # Disease Prognosis and Precautions Text2Text Generation Welcome to the Disease Prognosis and Precautions Text2Text Generation repository! Fine-tuned microsoft/GODEL-v1_1-large-seq2seq. The model is designed to generate responses for disease prognosis and recommended precautions based on given symptoms. ## Model Overview The model in this repository is a text-to-text generation model. It takes a prompt in the form of symptoms related to a particular disease and generates a response that includes the potential disease prognosis along with recommended precautions. The columns used in the training dataset are: - **Disease:** The name of the disease related to the symptoms. - **Symptoms:** The list of symptoms provided in the prompt. - **Precautions:** The recommended precautions for the identified disease. ## Examples Here are some examples of how you can use the model: ### Example 1 **Prompt:** "I am feeling continuous sneezing, shivering and chills" **Response:** "Seems like allergy. You should try to avoid dust and air pollution." ### Example 2 **Prompt:** "I am feeling itching, skin rash and patches" **Response:** "Seems like fungal infection. You should bathe twice a day and use antifungal soap." ## How to Use To use the model for generating disease prognosis and precautions based on symptoms, you can use the `generate` function provided by the Hugging Face Transformers library. Here's a basic example using Python: ```python from transformers import AutoModelForSeq2SeqLM, AutoTokenizer # Load the model and tokenizer model_name = "shanover/medbot_godel_v3" model = AutoModelForSeq2SeqLM.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) # Define your symptom prompt prompt = "I am feeling continuous sneezing, shivering and chills" def generate_response(input_text, model, tokenizer, max_length): input_ids = tokenizer.encode(input_text, return_tensors="pt", max_length=max_length, truncation=True) input_ids = input_ids.to(device) with torch.no_grad(): output_ids = model.generate(input_ids) generated_text = tokenizer.decode(output_ids[0], skip_special_tokens=True) return generated_text print(generate_response(prompt, model, tokenizer)) ``` Remember to replace `"shanover/medbot_godel_v3"` with the actual name or path of the model you've downloaded or fine-tuned. ## Acknowledgments Trained on Microsoft/Godel: https://huggingface.co/microsoft/GODEL-v1_1-large-seq2seq ## Issues and Contributions If you encounter any issues while using the model or have suggestions for improvements, please feel free to open an issue in this repository. Contributions are also welcome! ## Disclaimer Please note that the information generated by the model is for informational purposes only and should not be considered a substitute for professional medical advice. Always consult a medical professional for accurate diagnoses and treatments. Thank you for using the Disease Prognosis and Precautions Text2Text Generation model! We hope it proves to be a helpful tool.