--- title: ExplainAnything.AI emoji: 📚 colorFrom: blue colorTo: purple sdk: gradio sdk_version: "4.20.0" app_file: app.py pinned: true --- # ExplainAnything.AI **Track:** agent-demo-track **ExplainAnything.AI** is a multimodal agent that helps users understand any science topic — either by **asking a question** or **uploading an image or PDF**. It then builds an interactive explainer package with: - 🧠 Easy-to-understand explanation (via Mistral) - 🖼️ Auto-generated visual diagram (via Flux) - ❓ Quiz questions to test understanding (via Mistral) - 📄 Downloadable report summarizing everything --- ## 🚀 How It Works You have two options to get started: 1. **Ask a question**, e.g. *"How do solar panels work?"* 2. **Upload an image or PDF**, like a diagram or worksheet. The agent then: - Uses **Gemini Vision** (for image/PDF) or **your question** as input - Generates an explanation using **Mistral** - Creates a visual diagram with **Flux** - Generates quiz questions using **Mistral** - Compiles everything into a downloadable **learning report** --- ## 🚧 Build Status This Space is currently **building**, but all logic, tools, and functionality were submitted **before the deadline**. --- ## 🎥 Video Overview **Video Overview:** [Coming Soon – will be added post-deadline] *A short walkthrough of the app’s flow and learning experience will be uploaded here.* --- ## 🛠️ Tech Stack - **Mistral** – for science explanations + quiz generation - **Flux** – for diagram/image generation - **Gemini Vision** – for reading image and PDF content - **Gradio** – chat + upload interface - Manual orchestration (no MCP yet) --- ## 📘 Use Cases - 🧑‍🎓 Students learning STEM with visual + interactive help - 👩‍🏫 Teachers turning textbook pages into visual lessons - 🧠 Self-learners asking "why/how" questions and getting full reports --- ## 🧠 Future Plans - Integrate Hugging Face MCP for agent orchestration - Add TTS narration for accessibility - Generate downloadable PDF learning packs