--- license: mit language: - en base_model: - distilbert/distilbert-base-uncased --- # DistilBERT Sentiment Analysis Model This model is a fine-tuned version of **DistilBERT** for sentiment analysis on the **IMDb** dataset. It classifies movie reviews as either **positive** or **negative** based on the text content. ## Model Details - **Model Type**: DistilBERT (a smaller and faster variant of BERT) - **Task**: Sentiment Analysis - **Dataset**: IMDb dataset containing movie reviews with labels (positive/negative) - **Fine-Tuned On**: IMDb dataset ## Model Performance This model was fine-tuned on the IMDb dataset for sentiment classification, achieving good performance for binary sentiment classification tasks (positive/negative). ## Usage To use this model, you can load it from the Hugging Face Model Hub using the `transformers` library: ```python from transformers import pipeline # Load the model classifier = pipeline('sentiment-analysis', model='dorukan/distilbert-base-uncased-bert-finetuned-imdb') # Example usage result = classifier("This movie was amazing!") print(result) ``` This will output a sentiment prediction for the given text. ## License This model is licensed under the MIT License. For more information, see the LICENSE file. ## Acknowledgments - **DistilBERT**: A smaller version of BERT, created by the Hugging Face team. - **IMDb Dataset**: A collection of movie reviews used for sentiment classification, widely used in NLP tasks. You can find more details about the model at the [Hugging Face model page](https://huggingface.co/dorukan/distil-bert-imdb).