import os import streamlit as st import requests import webbrowser import time import base64 from google.generativeai import configure, GenerativeModel import streamlit.components.v1 as components from urllib.parse import urlparse from streamlit_lottie import st_lottie from together import Together import smtplib from email.mime.multipart import MIMEMultipart from email.mime.text import MIMEText from email.mime.base import MIMEBase from email import encoders from email.utils import formataddr from io import BytesIO from PIL import Image # Configure the Generative AI and Together clients configure(api_key="AIzaSyBOc7WOykXVHvnU-GsMgCYZwoBqFERjQFI") model = GenerativeModel('gemini-pro') together_api_key = "00cb552f0e0a189c90abd086094fc503d6a8091e35634a769bacdc4266a3af6f" client = Together(api_key=together_api_key) # Email sending function def send_email(subject, body, to_email, attachment=None, filename="attachment.png"): from_email = st.secrets["from_email"] from_password = st.secrets["from_password"] msg = MIMEMultipart() msg['From'] = formataddr(('AutoBot', from_email)) msg['To'] = to_email msg['Subject'] = subject msg.attach(MIMEText(body, 'plain')) if attachment: part = MIMEBase('application', 'octet-stream') part.set_payload(attachment) encoders.encode_base64(part) part.add_header('Content-Disposition', f'attachment; filename={filename}') msg.attach(part) try: server = smtplib.SMTP('smtp.gmail.com', 587) server.starttls() server.login(from_email, from_password) text = msg.as_string() server.sendmail(from_email, to_email, text) server.quit() st.success("Email sent successfully!") except Exception as e: st.error(f"Failed to send email: {e}") # Lottie animation loader def load_lottie_url(url: str): response = requests.get(url) if response.status_code != 200: return None return response.json() # Function to download HTML code def download_html_code(html_content, url): try: domain = urlparse(url).netloc.replace('www.', '') filename = f"{domain}_code.html" with open(filename, 'w') as file: file.write(html_content) st.markdown(get_binary_file_downloader_html(filename), unsafe_allow_html=True) except Exception as e: st.error(f"Failed to download HTML code: {e}") # Function to redirect to GitHub Codespaces def redirect_to_codespaces(): with st.spinner("Redirecting to GitHub Codespaces..."): time.sleep(2) webbrowser.open_new_tab("https://github.com/codespaces") st.info("If the application can't redirect, use the link below:") st.markdown("[GitHub Codespaces](https://github.com/codespaces)") # Function to download generated code def download_generated_code(content, filename, format='txt'): extension = format temp_filename = f"{filename}.{extension}" with open(temp_filename, 'w') as file: file.write(content) with open(temp_filename, 'rb') as file: data = file.read() b64 = base64.b64encode(data).decode() href = f'Download Code ({format.upper()})' st.markdown(href, unsafe_allow_html=True) os.remove(temp_filename) # Function to display file download link def get_binary_file_downloader_html(bin_file, file_label='Download Code'): with open(bin_file, 'rb') as f: data = f.read() b64 = base64.b64encode(data).decode() href = f'{file_label}' return href # Function to display footer def display_footer(): footer_html = """ """ st.markdown(footer_html, unsafe_allow_html=True) # Function to fetch YouTube video suggestions def fetch_youtube_videos(query): api_key = st.secrets["youtube_api_key"] search_url = "https://www.googleapis.com/youtube/v3/search" params = { "part": "snippet", "q": query, "type": "video", "maxResults": 4, "key": api_key } response = requests.get(search_url, params=params) video_details = [] if response.status_code == 200: results = response.json()["items"] for item in results: video_id = item["id"]["videoId"] video_title = item["snippet"]["title"] video_url = f"https://www.youtube.com/watch?v={video_id}" video_details.append({ "title": video_title, "url": video_url, "video_id": video_id }) else: st.error(f"Failed to fetch YouTube videos. Status code: {response.status_code}") return video_details # Function to extract the main topic from the prompt def extract_topic(prompt): start_phrases = ["@codex", "codex", "@autobot"] for phrase in start_phrases: if prompt.lower().startswith(phrase): return prompt[len(phrase):].strip() return prompt.strip() # Main Streamlit application def main(): st.set_page_config(page_title="AutoBot AI", page_icon="💀", layout="wide", initial_sidebar_state="expanded") st.sidebar.image("autobot2.png", use_column_width=True) page = st.sidebar.selectbox("**MENU**", ["🏠 Home", "AutoBot 💀", "CODEX ⚡", "Web Scrapper 🌐", "GitHub Codespaces 🖥️", "Mega Bot 🐸", "Refund & Privacy Policy 💸", ]) st.sidebar.title("Support Us") st.sidebar.info("Your support helps us improve AutoBot AI.") if page == "🏠 Home": st.title("Welcome to AutoBot AI 💀") st.markdown(""" **AutoBot AI**: **Functionalities:** 1. AI Chatbot 2. CODEX 3. Web Scrapper 4. GitHub Codespaces AutoBot, powered by the Gemini API, is a basic chatbot designed for automation. It excels in writing code and generating downloadable files with a .txt extension, offering the ability to handle up to 60 queries per minute. Developed by SKAV TECH, a company focused on creating practical AI projects, AutoBot is intended for educational purposes only. We do not endorse any illegal or unethical activities. """) # Embedding Lottie animation st.markdown(""" """, unsafe_allow_html=True) st.video("https://youtu.be/i0Q-NBrYpPI", start_time=0) display_footer() st.sidebar.markdown("If you find this tool helpful, consider supporting us.") components.html( """
""", height=450, width=300 ) elif page == "AutoBot 💀": st.image("autobot2.png") st.header("AutoBot 💀") st.markdown( "AutoBot is effective for code generation. If your prompt contains code generation **-prompt-**, you can get downloadable files.") question = st.text_input("Ask the model a question:") if st.button("Ask AI"): lottie_url = "https://lottie.host/fb24aa71-e6dd-497e-8a6c-3098cb64b1ed/V9N1Sd3klS.json" # Load and display Lottie animation lottie_animation = load_lottie_url(lottie_url) if lottie_animation: st_lottie(lottie_animation, speed=27, width=150, height=100, key="lottie_animation") else: st.error("Failed to load Lottie animation.") with st.spinner("Generating response 💀..."): try: response = model.generate_content(question) if response.text: st.text("AutoBot Response:") st.write(response.text) st.markdown('---') st.markdown( "Security Note: We use **.txt** file format for code downloads, which is not easily susceptible to virus and malware attacks.") else: st.error("No valid response received from the AI model.") st.write(f"Safety ratings: {response.safety_ratings}") except ValueError as e: st.info(f"Unable to assist with that prompt due to: {e}") except IndexError as e: st.info(f"Unable to assist with that prompt due to: {e}") except Exception as e: st.info(f"An unexpected error occurred: {e}") code_keywords = ["code", "write code", "develop code", "generate code", "generate", "build"] if any(keyword in question.lower() for keyword in code_keywords): st.text("Download the generated code 💀:") download_generated_code(response.text, "code", format='txt') st.sidebar.markdown("If you find this tool helpful, consider supporting us.") components.html( """
""", height=450, width=300 ) display_footer() elif page == "CODEX ⚡": st.image("codex.png") st.header("CODEX ⚡") st.markdown( "The CODEX tool is effective for code generation and provides YouTube video suggestions based on your code-related queries.") question = st.text_input("Ask the model for code generation:") if st.button("Generate Code"): with st.spinner("Generating code..."): try: response = model.generate_content(question) if response.text: st.text("CODEX Response:") st.write(response.text) download_generated_code(response.text, "code", format='txt') else: st.error("No valid response received from the AI model.") except ValueError as e: st.info(f"Unable to assist with that prompt due to: {e}") except IndexError as e: st.info(f"Unable to assist with that prompt due to: {e}") except Exception as e: st.info(f"An unexpected error occurred: {e}") topic = extract_topic(question) video_results = fetch_youtube_videos(topic) if video_results: st.subheader("YouTube Video Suggestions") for video in video_results: st.markdown(f"[{video['title']}]({video['url']})") st.video(video['url']) else: st.info("No relevant YouTube videos found.") st.sidebar.markdown("If you find this tool helpful, consider supporting us.") components.html( """
""", height=450, width=300 ) display_footer() elif page == "Web Scrapper 🌐": st.image("autobot2.png") st.header("Web Scrapper 🌐") url = st.text_input("Enter URL to scrape:") if st.button("Scrape HTML Code"): if url: with st.spinner("Scraping HTML code..."): try: response = requests.get(url) response.raise_for_status() html_content = response.text st.code(html_content, language="html") download_html_code(html_content, url) except requests.exceptions.RequestException as e: st.error(f"Failed to scrape HTML code: {e}") else: st.error("Please enter a valid URL.") st.sidebar.markdown("If you find this tool helpful, consider supporting us.") components.html( """
""", height=450, width=300 ) display_footer() elif page == "GitHub Codespaces 🖥️": st.image("autobot2.png") st.header("GitHub Codespaces 🖥️") if st.button("Open GitHub Codespaces"): redirect_to_codespaces() display_footer() elif page == "Refund & Privacy Policy 💸": st.image("autobot2.png") st.header("Refund & Privacy Policy 💸") st.markdown( """ ## Refund Policy We want you to be satisfied with our services. If you have any issues or concerns, please contact us within 30 days of purchase. We will review your request and provide a refund if deemed appropriate. ## Privacy Policy Your privacy is important to us. We collect only the necessary data to provide our services and do not share your information with third parties without your consent. Please review our [full privacy policy](https://www.example.com/privacy-policy) for more details. """ ) display_footer() elif page == "Mega Bot 🐸": st.image("megabot.png") st.markdown('---') st.subheader("🤖 Multi-model AI Application") st.markdown( "This application integrates multiple AI models and tools for various functionalities such as chat, code generation, image generation.") # Tabs for navigation tabs = st.tabs(["General Chat", "Code Generation", "Image Generation"]) # Load animations chat_animation = load_lottie_url("https://assets2.lottiefiles.com/private_files/lf30_xTmPwn.json") code_animation = load_lottie_url("https://assets5.lottiefiles.com/packages/lf20_ba55esn2.json") image_animation = load_lottie_url("https://assets2.lottiefiles.com/private_files/lf30_O5QGL0.json") # General Chat Tab with tabs[0]: st.header("💬 General Chat") user_prompt = st.text_input("Enter your prompt:", "") if st.button("Generate Response"): with st.spinner("Generating response..."): response = model.generate_content(user_prompt) st.markdown('---') st.write(response.text) st.markdown('---') st.sidebar.write(""" **General Chat Instructions:** 1. Enter your prompt in the text box. 2. Click on the 'Generate Response' button to see the AI's response. """) st.sidebar.markdown("If you find this tool helpful, consider supporting us.") components.html( """
""", height=450, width=300 ) # Code Generation Tab with tabs[1]: st.header("💻 Code Generation") user_prompt = st.text_input("Enter your coding prompt:", "") if st.button("Generate Code"): with st.spinner("Generating code..."): response = client.chat.completions.create( model="codellama/CodeLlama-70b-Instruct-hf", messages=[{"role": "user", "content": user_prompt}], ) st.code(response.choices[0].message.content, language="python") st.sidebar.write(""" **Code Generation Instructions:** 1. Enter your coding prompt in the text box. 2. Click on the 'Generate Code' button to see the generated code. """) # Image Generation Tab with tabs[2]: st.header("🖼️ Image Generation") st.markdown('---') st.info( "We are currently working on Tuning the Models, So the AI generated images might not match your prompts. Improve your Prompt Context to get good results 😊") st.markdown('---') user_prompt = st.text_input("Enter your image prompt:", "") model_choice = st.selectbox("Choose the image model", [ "SG161222/Realistic_Vision_V3.0_VAE", "stabilityai/stable-diffusion-xl-base-1.0", "stabilityai/stable-diffusion-2-1", "runwayml/stable-diffusion-v1-5", "prompthero/openjourney" ]) if st.button("Generate Image"): with st.spinner("Generating image..."): response = client.images.generate( prompt=user_prompt, model=model_choice, steps=10, n=1 ) img_data = response.data[0].b64_json img_bytes = base64.b64decode(img_data) img = Image.open(BytesIO(img_bytes)) st.image(img) email_subject = "New Image Generation Prompt" email_body = f"Prompt: {user_prompt}\nModel: {model_choice}\nDetails: Your additional details here." send_email(email_subject, email_body, "skavtech.in@gmail.com", attachment=img_bytes, filename="generated_image.png") st.sidebar.write(""" **Image Generation Instructions:** 1. Enter your image prompt in the text box. 2. Choose an image generation model from the dropdown. 3. Click on the 'Generate Image' button to see the generated image. """) st.sidebar.markdown("If you find this tool helpful, consider supporting us.") components.html( """
""", height=450, width=300 ) # Run the app with: streamlit run app.py display_footer() if __name__ == "__main__": main()