Spaces:
				
			
			
	
			
			
		Sleeping
		
	
	
	
			
			
	
	
	
	
		
		
		Sleeping
		
	| import gradio as gr | |
| import tensorflow as tf | |
| import pickle | |
| import numpy as np | |
| import requests | |
| from ProGPT import Conversation | |
| # Load saved components | |
| with open('preprocessing_params.pkl', 'rb') as f: | |
| preprocessing_params = pickle.load(f) | |
| with open('fisher_information.pkl', 'rb') as f: | |
| fisher_information = pickle.load(f) | |
| with open('label_encoder.pkl', 'rb') as f: | |
| label_encoder = pickle.load(f) | |
| with open('url_tokenizer.pkl', 'rb') as f: | |
| url_tokenizer = pickle.load(f) | |
| with open('html_tokenizer.pkl', 'rb') as f: | |
| html_tokenizer = pickle.load(f) | |
| # Load the model with custom loss | |
| class EWCLoss(tf.keras.losses.Loss): | |
| def __init__(self, model, fisher_information, importance=1.0, reduction='auto', name=None): | |
| super(EWCLoss, self).__init__(reduction=reduction, name=name) | |
| self.model = model | |
| self.fisher_information = fisher_information | |
| self.importance = importance | |
| self.prev_weights = [layer.numpy() for layer in model.trainable_weights] | |
| def call(self, y_true, y_pred): | |
| standard_loss = tf.keras.losses.binary_crossentropy(y_true, y_pred) | |
| ewc_loss = 0.0 | |
| for layer, fisher_info, prev_weight in zip(self.model.trainable_weights, self.fisher_information, self.prev_weights): | |
| ewc_loss += tf.reduce_sum(fisher_info * tf.square(layer - prev_weight)) | |
| return standard_loss + (self.importance / 2.0) * ewc_loss | |
| def get_config(self): | |
| config = super().get_config() | |
| config.update({ | |
| 'importance': self.importance, | |
| 'reduction': self.reduction, | |
| 'name': self.name, | |
| }) | |
| return config | |
| def from_config(cls, config): | |
| # Load fisher information from external file | |
| with open('fisher_information.pkl', 'rb') as f: | |
| fisher_information = pickle.load(f) | |
| return cls(model=None, fisher_information=fisher_information, **config) | |
| # Load the model | |
| model = tf.keras.models.load_model('new_phishing_detection_model.keras', | |
| custom_objects={'EWCLoss': EWCLoss}) | |
| # Recompile the model | |
| ewc_loss = EWCLoss(model=model, fisher_information=fisher_information, importance=1000) | |
| model.compile(optimizer=tf.keras.optimizers.Adam(learning_rate=0.0005), | |
| loss=ewc_loss, | |
| metrics=['accuracy', tf.keras.metrics.Precision(), tf.keras.metrics.Recall()]) | |
| # Chatbot setup | |
| access_token = 'your_pro_gpt_access_token' | |
| chatbot = Conversation(access_token) | |
| # Function to preprocess input | |
| def preprocess_input(input_text, tokenizer, max_length): | |
| sequences = tokenizer.texts_to_sequences([input_text]) | |
| padded_sequences = tf.keras.preprocessing.sequence.pad_sequences(sequences, maxlen=max_length, padding='post', truncating='post') | |
| return padded_sequences | |
| # Function to get prediction | |
| def get_prediction(input_text, input_type): | |
| is_url = input_type == "URL" | |
| if is_url: | |
| input_data = preprocess_input(input_text, url_tokenizer, preprocessing_params['max_url_length']) | |
| else: | |
| input_data = preprocess_input(input_text, html_tokenizer, preprocessing_params['max_html_length']) | |
| prediction = model.predict([input_data, input_data])[0][0] | |
| return prediction | |
| # Function to fetch latest phishing sites from PhishTank | |
| def fetch_latest_phishing_sites(): | |
| try: | |
| response = requests.get('https://data.phishtank.com/data/online-valid.json') | |
| data = response.json() | |
| return data[:5] | |
| except Exception as e: | |
| return [] | |
| # Gradio UI | |
| def phishing_detection(input_text, input_type): | |
| prediction = get_prediction(input_text, input_type) | |
| if prediction > 0.5: | |
| return f"Warning: This site is likely a phishing site! ({prediction:.2f})" | |
| else: | |
| return f"Safe: This site is not likely a phishing site. ({prediction:.2f})" | |
| def latest_phishing_sites(): | |
| sites = fetch_latest_phishing_sites() | |
| return [f"{site['url']}" for site in sites] | |
| def chatbot_response(user_input): | |
| response = chatbot.prompt(user_input) | |
| return response | |
| iface = gr.Interface( | |
| fn=phishing_detection, | |
| inputs=[gr.inputs.Textbox(lines=5, placeholder="Enter URL or HTML code"), gr.inputs.Radio(["URL", "HTML"], type="value", label="Input Type")], | |
| outputs="text", | |
| title="Phishing Detection with Enhanced EWC Model", | |
| description="Check if a URL or HTML is Phishing", | |
| theme="default" | |
| ) | |
| iface.launch() | 
