{ "cells": [ { "cell_type": "code", "execution_count": 1, "metadata": {}, "outputs": [ { "ename": "NameError", "evalue": "name 'create_repo' is not defined", "output_type": "error", "traceback": [ "\u001b[31m---------------------------------------------------------------------------\u001b[39m", "\u001b[31mNameError\u001b[39m Traceback (most recent call last)", "\u001b[36mCell\u001b[39m\u001b[36m \u001b[39m\u001b[32mIn[1]\u001b[39m\u001b[32m, line 2\u001b[39m\n\u001b[32m 1\u001b[39m repo_id = \u001b[33m\"\u001b[39m\u001b[33meunJ/codebert_vulnerability_detector_multi\u001b[39m\u001b[33m\"\u001b[39m \u001b[38;5;66;03m# 바꿔주세요\u001b[39;00m\n\u001b[32m----> \u001b[39m\u001b[32m2\u001b[39m \u001b[43mcreate_repo\u001b[49m(repo_id, repo_type=\u001b[33m\"\u001b[39m\u001b[33mmodel\u001b[39m\u001b[33m\"\u001b[39m, exist_ok=\u001b[38;5;28;01mTrue\u001b[39;00m)\n", "\u001b[31mNameError\u001b[39m: name 'create_repo' is not defined" ] } ], "source": [ "repo_id = \"eunJ/codebert_vulnerability_detector_multi\" # 바꿔주세요\n", "create_repo(repo_id, repo_type=\"model\", exist_ok=True)\n" ] }, { "cell_type": "code", "execution_count": 2, "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "/data/eun/anaconda3/envs/xcodevul/lib/python3.13/site-packages/tqdm/auto.py:21: TqdmWarning: IProgress not found. Please update jupyter and ipywidgets. See https://ipywidgets.readthedocs.io/en/stable/user_install.html\n", " from .autonotebook import tqdm as notebook_tqdm\n" ] } ], "source": [ "\n", "from huggingface_hub import create_repo, upload_folder\n", "import os" ] }, { "cell_type": "code", "execution_count": 5, "metadata": {}, "outputs": [ { "data": { "text/plain": [ "CommitInfo(commit_url='https://huggingface.co/eunJ/codebert_vulnerability_detector_multi/commit/71f5c4ac3fbe8a56ff25c7bd1c69825246eb264f', commit_message='Initial commit: upload all files in current directory', commit_description='', oid='71f5c4ac3fbe8a56ff25c7bd1c69825246eb264f', pr_url=None, repo_url=RepoUrl('https://huggingface.co/eunJ/codebert_vulnerability_detector_multi', endpoint='https://huggingface.co', repo_type='model', repo_id='eunJ/codebert_vulnerability_detector_multi'), pr_revision=None, pr_num=None)" ] }, "execution_count": 5, "metadata": {}, "output_type": "execute_result" } ], "source": [ "local_dir = \".\" # 현재 디렉토리 전체\n", "\n", "# 폴더 전체 업로드\n", "upload_folder(\n", " folder_path=local_dir,\n", " repo_id=repo_id,\n", " repo_type=\"model\",\n", " commit_message=\"Initial commit: upload all files in current directory\"\n", ")" ] }, { "cell_type": "code", "execution_count": 4, "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "Some weights of the model checkpoint at eunJ/codebert_vulnerability_detector_multi were not used when initializing RobertaForSequenceClassification: ['classifier.bias', 'classifier.weight', 'encoder.classifier.dense.bias', 'encoder.classifier.dense.weight', 'encoder.classifier.out_proj.bias', 'encoder.classifier.out_proj.weight', 'encoder.roberta.embeddings.LayerNorm.bias', 'encoder.roberta.embeddings.LayerNorm.weight', 'encoder.roberta.embeddings.position_embeddings.weight', 'encoder.roberta.embeddings.token_type_embeddings.weight', 'encoder.roberta.embeddings.word_embeddings.weight', 'encoder.roberta.encoder.layer.0.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.0.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.0.attention.output.dense.bias', 'encoder.roberta.encoder.layer.0.attention.output.dense.weight', 'encoder.roberta.encoder.layer.0.attention.self.key.bias', 'encoder.roberta.encoder.layer.0.attention.self.key.weight', 'encoder.roberta.encoder.layer.0.attention.self.query.bias', 'encoder.roberta.encoder.layer.0.attention.self.query.weight', 'encoder.roberta.encoder.layer.0.attention.self.value.bias', 'encoder.roberta.encoder.layer.0.attention.self.value.weight', 'encoder.roberta.encoder.layer.0.intermediate.dense.bias', 'encoder.roberta.encoder.layer.0.intermediate.dense.weight', 'encoder.roberta.encoder.layer.0.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.0.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.0.output.dense.bias', 'encoder.roberta.encoder.layer.0.output.dense.weight', 'encoder.roberta.encoder.layer.1.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.1.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.1.attention.output.dense.bias', 'encoder.roberta.encoder.layer.1.attention.output.dense.weight', 'encoder.roberta.encoder.layer.1.attention.self.key.bias', 'encoder.roberta.encoder.layer.1.attention.self.key.weight', 'encoder.roberta.encoder.layer.1.attention.self.query.bias', 'encoder.roberta.encoder.layer.1.attention.self.query.weight', 'encoder.roberta.encoder.layer.1.attention.self.value.bias', 'encoder.roberta.encoder.layer.1.attention.self.value.weight', 'encoder.roberta.encoder.layer.1.intermediate.dense.bias', 'encoder.roberta.encoder.layer.1.intermediate.dense.weight', 'encoder.roberta.encoder.layer.1.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.1.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.1.output.dense.bias', 'encoder.roberta.encoder.layer.1.output.dense.weight', 'encoder.roberta.encoder.layer.10.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.10.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.10.attention.output.dense.bias', 'encoder.roberta.encoder.layer.10.attention.output.dense.weight', 'encoder.roberta.encoder.layer.10.attention.self.key.bias', 'encoder.roberta.encoder.layer.10.attention.self.key.weight', 'encoder.roberta.encoder.layer.10.attention.self.query.bias', 'encoder.roberta.encoder.layer.10.attention.self.query.weight', 'encoder.roberta.encoder.layer.10.attention.self.value.bias', 'encoder.roberta.encoder.layer.10.attention.self.value.weight', 'encoder.roberta.encoder.layer.10.intermediate.dense.bias', 'encoder.roberta.encoder.layer.10.intermediate.dense.weight', 'encoder.roberta.encoder.layer.10.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.10.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.10.output.dense.bias', 'encoder.roberta.encoder.layer.10.output.dense.weight', 'encoder.roberta.encoder.layer.11.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.11.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.11.attention.output.dense.bias', 'encoder.roberta.encoder.layer.11.attention.output.dense.weight', 'encoder.roberta.encoder.layer.11.attention.self.key.bias', 'encoder.roberta.encoder.layer.11.attention.self.key.weight', 'encoder.roberta.encoder.layer.11.attention.self.query.bias', 'encoder.roberta.encoder.layer.11.attention.self.query.weight', 'encoder.roberta.encoder.layer.11.attention.self.value.bias', 'encoder.roberta.encoder.layer.11.attention.self.value.weight', 'encoder.roberta.encoder.layer.11.intermediate.dense.bias', 'encoder.roberta.encoder.layer.11.intermediate.dense.weight', 'encoder.roberta.encoder.layer.11.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.11.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.11.output.dense.bias', 'encoder.roberta.encoder.layer.11.output.dense.weight', 'encoder.roberta.encoder.layer.2.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.2.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.2.attention.output.dense.bias', 'encoder.roberta.encoder.layer.2.attention.output.dense.weight', 'encoder.roberta.encoder.layer.2.attention.self.key.bias', 'encoder.roberta.encoder.layer.2.attention.self.key.weight', 'encoder.roberta.encoder.layer.2.attention.self.query.bias', 'encoder.roberta.encoder.layer.2.attention.self.query.weight', 'encoder.roberta.encoder.layer.2.attention.self.value.bias', 'encoder.roberta.encoder.layer.2.attention.self.value.weight', 'encoder.roberta.encoder.layer.2.intermediate.dense.bias', 'encoder.roberta.encoder.layer.2.intermediate.dense.weight', 'encoder.roberta.encoder.layer.2.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.2.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.2.output.dense.bias', 'encoder.roberta.encoder.layer.2.output.dense.weight', 'encoder.roberta.encoder.layer.3.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.3.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.3.attention.output.dense.bias', 'encoder.roberta.encoder.layer.3.attention.output.dense.weight', 'encoder.roberta.encoder.layer.3.attention.self.key.bias', 'encoder.roberta.encoder.layer.3.attention.self.key.weight', 'encoder.roberta.encoder.layer.3.attention.self.query.bias', 'encoder.roberta.encoder.layer.3.attention.self.query.weight', 'encoder.roberta.encoder.layer.3.attention.self.value.bias', 'encoder.roberta.encoder.layer.3.attention.self.value.weight', 'encoder.roberta.encoder.layer.3.intermediate.dense.bias', 'encoder.roberta.encoder.layer.3.intermediate.dense.weight', 'encoder.roberta.encoder.layer.3.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.3.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.3.output.dense.bias', 'encoder.roberta.encoder.layer.3.output.dense.weight', 'encoder.roberta.encoder.layer.4.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.4.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.4.attention.output.dense.bias', 'encoder.roberta.encoder.layer.4.attention.output.dense.weight', 'encoder.roberta.encoder.layer.4.attention.self.key.bias', 'encoder.roberta.encoder.layer.4.attention.self.key.weight', 'encoder.roberta.encoder.layer.4.attention.self.query.bias', 'encoder.roberta.encoder.layer.4.attention.self.query.weight', 'encoder.roberta.encoder.layer.4.attention.self.value.bias', 'encoder.roberta.encoder.layer.4.attention.self.value.weight', 'encoder.roberta.encoder.layer.4.intermediate.dense.bias', 'encoder.roberta.encoder.layer.4.intermediate.dense.weight', 'encoder.roberta.encoder.layer.4.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.4.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.4.output.dense.bias', 'encoder.roberta.encoder.layer.4.output.dense.weight', 'encoder.roberta.encoder.layer.5.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.5.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.5.attention.output.dense.bias', 'encoder.roberta.encoder.layer.5.attention.output.dense.weight', 'encoder.roberta.encoder.layer.5.attention.self.key.bias', 'encoder.roberta.encoder.layer.5.attention.self.key.weight', 'encoder.roberta.encoder.layer.5.attention.self.query.bias', 'encoder.roberta.encoder.layer.5.attention.self.query.weight', 'encoder.roberta.encoder.layer.5.attention.self.value.bias', 'encoder.roberta.encoder.layer.5.attention.self.value.weight', 'encoder.roberta.encoder.layer.5.intermediate.dense.bias', 'encoder.roberta.encoder.layer.5.intermediate.dense.weight', 'encoder.roberta.encoder.layer.5.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.5.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.5.output.dense.bias', 'encoder.roberta.encoder.layer.5.output.dense.weight', 'encoder.roberta.encoder.layer.6.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.6.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.6.attention.output.dense.bias', 'encoder.roberta.encoder.layer.6.attention.output.dense.weight', 'encoder.roberta.encoder.layer.6.attention.self.key.bias', 'encoder.roberta.encoder.layer.6.attention.self.key.weight', 'encoder.roberta.encoder.layer.6.attention.self.query.bias', 'encoder.roberta.encoder.layer.6.attention.self.query.weight', 'encoder.roberta.encoder.layer.6.attention.self.value.bias', 'encoder.roberta.encoder.layer.6.attention.self.value.weight', 'encoder.roberta.encoder.layer.6.intermediate.dense.bias', 'encoder.roberta.encoder.layer.6.intermediate.dense.weight', 'encoder.roberta.encoder.layer.6.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.6.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.6.output.dense.bias', 'encoder.roberta.encoder.layer.6.output.dense.weight', 'encoder.roberta.encoder.layer.7.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.7.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.7.attention.output.dense.bias', 'encoder.roberta.encoder.layer.7.attention.output.dense.weight', 'encoder.roberta.encoder.layer.7.attention.self.key.bias', 'encoder.roberta.encoder.layer.7.attention.self.key.weight', 'encoder.roberta.encoder.layer.7.attention.self.query.bias', 'encoder.roberta.encoder.layer.7.attention.self.query.weight', 'encoder.roberta.encoder.layer.7.attention.self.value.bias', 'encoder.roberta.encoder.layer.7.attention.self.value.weight', 'encoder.roberta.encoder.layer.7.intermediate.dense.bias', 'encoder.roberta.encoder.layer.7.intermediate.dense.weight', 'encoder.roberta.encoder.layer.7.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.7.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.7.output.dense.bias', 'encoder.roberta.encoder.layer.7.output.dense.weight', 'encoder.roberta.encoder.layer.8.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.8.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.8.attention.output.dense.bias', 'encoder.roberta.encoder.layer.8.attention.output.dense.weight', 'encoder.roberta.encoder.layer.8.attention.self.key.bias', 'encoder.roberta.encoder.layer.8.attention.self.key.weight', 'encoder.roberta.encoder.layer.8.attention.self.query.bias', 'encoder.roberta.encoder.layer.8.attention.self.query.weight', 'encoder.roberta.encoder.layer.8.attention.self.value.bias', 'encoder.roberta.encoder.layer.8.attention.self.value.weight', 'encoder.roberta.encoder.layer.8.intermediate.dense.bias', 'encoder.roberta.encoder.layer.8.intermediate.dense.weight', 'encoder.roberta.encoder.layer.8.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.8.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.8.output.dense.bias', 'encoder.roberta.encoder.layer.8.output.dense.weight', 'encoder.roberta.encoder.layer.9.attention.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.9.attention.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.9.attention.output.dense.bias', 'encoder.roberta.encoder.layer.9.attention.output.dense.weight', 'encoder.roberta.encoder.layer.9.attention.self.key.bias', 'encoder.roberta.encoder.layer.9.attention.self.key.weight', 'encoder.roberta.encoder.layer.9.attention.self.query.bias', 'encoder.roberta.encoder.layer.9.attention.self.query.weight', 'encoder.roberta.encoder.layer.9.attention.self.value.bias', 'encoder.roberta.encoder.layer.9.attention.self.value.weight', 'encoder.roberta.encoder.layer.9.intermediate.dense.bias', 'encoder.roberta.encoder.layer.9.intermediate.dense.weight', 'encoder.roberta.encoder.layer.9.output.LayerNorm.bias', 'encoder.roberta.encoder.layer.9.output.LayerNorm.weight', 'encoder.roberta.encoder.layer.9.output.dense.bias', 'encoder.roberta.encoder.layer.9.output.dense.weight']\n", "- This IS expected if you are initializing RobertaForSequenceClassification from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).\n", "- This IS NOT expected if you are initializing RobertaForSequenceClassification from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).\n", "Some weights of RobertaForSequenceClassification were not initialized from the model checkpoint at eunJ/codebert_vulnerability_detector_multi and are newly initialized: ['classifier.dense.bias', 'classifier.dense.weight', 'classifier.out_proj.bias', 'classifier.out_proj.weight', 'embeddings.LayerNorm.bias', 'embeddings.LayerNorm.weight', 'embeddings.position_embeddings.weight', 'embeddings.token_type_embeddings.weight', 'embeddings.word_embeddings.weight', 'encoder.layer.0.attention.output.LayerNorm.bias', 'encoder.layer.0.attention.output.LayerNorm.weight', 'encoder.layer.0.attention.output.dense.bias', 'encoder.layer.0.attention.output.dense.weight', 'encoder.layer.0.attention.self.key.bias', 'encoder.layer.0.attention.self.key.weight', 'encoder.layer.0.attention.self.query.bias', 'encoder.layer.0.attention.self.query.weight', 'encoder.layer.0.attention.self.value.bias', 'encoder.layer.0.attention.self.value.weight', 'encoder.layer.0.intermediate.dense.bias', 'encoder.layer.0.intermediate.dense.weight', 'encoder.layer.0.output.LayerNorm.bias', 'encoder.layer.0.output.LayerNorm.weight', 'encoder.layer.0.output.dense.bias', 'encoder.layer.0.output.dense.weight', 'encoder.layer.1.attention.output.LayerNorm.bias', 'encoder.layer.1.attention.output.LayerNorm.weight', 'encoder.layer.1.attention.output.dense.bias', 'encoder.layer.1.attention.output.dense.weight', 'encoder.layer.1.attention.self.key.bias', 'encoder.layer.1.attention.self.key.weight', 'encoder.layer.1.attention.self.query.bias', 'encoder.layer.1.attention.self.query.weight', 'encoder.layer.1.attention.self.value.bias', 'encoder.layer.1.attention.self.value.weight', 'encoder.layer.1.intermediate.dense.bias', 'encoder.layer.1.intermediate.dense.weight', 'encoder.layer.1.output.LayerNorm.bias', 'encoder.layer.1.output.LayerNorm.weight', 'encoder.layer.1.output.dense.bias', 'encoder.layer.1.output.dense.weight', 'encoder.layer.10.attention.output.LayerNorm.bias', 'encoder.layer.10.attention.output.LayerNorm.weight', 'encoder.layer.10.attention.output.dense.bias', 'encoder.layer.10.attention.output.dense.weight', 'encoder.layer.10.attention.self.key.bias', 'encoder.layer.10.attention.self.key.weight', 'encoder.layer.10.attention.self.query.bias', 'encoder.layer.10.attention.self.query.weight', 'encoder.layer.10.attention.self.value.bias', 'encoder.layer.10.attention.self.value.weight', 'encoder.layer.10.intermediate.dense.bias', 'encoder.layer.10.intermediate.dense.weight', 'encoder.layer.10.output.LayerNorm.bias', 'encoder.layer.10.output.LayerNorm.weight', 'encoder.layer.10.output.dense.bias', 'encoder.layer.10.output.dense.weight', 'encoder.layer.11.attention.output.LayerNorm.bias', 'encoder.layer.11.attention.output.LayerNorm.weight', 'encoder.layer.11.attention.output.dense.bias', 'encoder.layer.11.attention.output.dense.weight', 'encoder.layer.11.attention.self.key.bias', 'encoder.layer.11.attention.self.key.weight', 'encoder.layer.11.attention.self.query.bias', 'encoder.layer.11.attention.self.query.weight', 'encoder.layer.11.attention.self.value.bias', 'encoder.layer.11.attention.self.value.weight', 'encoder.layer.11.intermediate.dense.bias', 'encoder.layer.11.intermediate.dense.weight', 'encoder.layer.11.output.LayerNorm.bias', 'encoder.layer.11.output.LayerNorm.weight', 'encoder.layer.11.output.dense.bias', 'encoder.layer.11.output.dense.weight', 'encoder.layer.2.attention.output.LayerNorm.bias', 'encoder.layer.2.attention.output.LayerNorm.weight', 'encoder.layer.2.attention.output.dense.bias', 'encoder.layer.2.attention.output.dense.weight', 'encoder.layer.2.attention.self.key.bias', 'encoder.layer.2.attention.self.key.weight', 'encoder.layer.2.attention.self.query.bias', 'encoder.layer.2.attention.self.query.weight', 'encoder.layer.2.attention.self.value.bias', 'encoder.layer.2.attention.self.value.weight', 'encoder.layer.2.intermediate.dense.bias', 'encoder.layer.2.intermediate.dense.weight', 'encoder.layer.2.output.LayerNorm.bias', 'encoder.layer.2.output.LayerNorm.weight', 'encoder.layer.2.output.dense.bias', 'encoder.layer.2.output.dense.weight', 'encoder.layer.3.attention.output.LayerNorm.bias', 'encoder.layer.3.attention.output.LayerNorm.weight', 'encoder.layer.3.attention.output.dense.bias', 'encoder.layer.3.attention.output.dense.weight', 'encoder.layer.3.attention.self.key.bias', 'encoder.layer.3.attention.self.key.weight', 'encoder.layer.3.attention.self.query.bias', 'encoder.layer.3.attention.self.query.weight', 'encoder.layer.3.attention.self.value.bias', 'encoder.layer.3.attention.self.value.weight', 'encoder.layer.3.intermediate.dense.bias', 'encoder.layer.3.intermediate.dense.weight', 'encoder.layer.3.output.LayerNorm.bias', 'encoder.layer.3.output.LayerNorm.weight', 'encoder.layer.3.output.dense.bias', 'encoder.layer.3.output.dense.weight', 'encoder.layer.4.attention.output.LayerNorm.bias', 'encoder.layer.4.attention.output.LayerNorm.weight', 'encoder.layer.4.attention.output.dense.bias', 'encoder.layer.4.attention.output.dense.weight', 'encoder.layer.4.attention.self.key.bias', 'encoder.layer.4.attention.self.key.weight', 'encoder.layer.4.attention.self.query.bias', 'encoder.layer.4.attention.self.query.weight', 'encoder.layer.4.attention.self.value.bias', 'encoder.layer.4.attention.self.value.weight', 'encoder.layer.4.intermediate.dense.bias', 'encoder.layer.4.intermediate.dense.weight', 'encoder.layer.4.output.LayerNorm.bias', 'encoder.layer.4.output.LayerNorm.weight', 'encoder.layer.4.output.dense.bias', 'encoder.layer.4.output.dense.weight', 'encoder.layer.5.attention.output.LayerNorm.bias', 'encoder.layer.5.attention.output.LayerNorm.weight', 'encoder.layer.5.attention.output.dense.bias', 'encoder.layer.5.attention.output.dense.weight', 'encoder.layer.5.attention.self.key.bias', 'encoder.layer.5.attention.self.key.weight', 'encoder.layer.5.attention.self.query.bias', 'encoder.layer.5.attention.self.query.weight', 'encoder.layer.5.attention.self.value.bias', 'encoder.layer.5.attention.self.value.weight', 'encoder.layer.5.intermediate.dense.bias', 'encoder.layer.5.intermediate.dense.weight', 'encoder.layer.5.output.LayerNorm.bias', 'encoder.layer.5.output.LayerNorm.weight', 'encoder.layer.5.output.dense.bias', 'encoder.layer.5.output.dense.weight', 'encoder.layer.6.attention.output.LayerNorm.bias', 'encoder.layer.6.attention.output.LayerNorm.weight', 'encoder.layer.6.attention.output.dense.bias', 'encoder.layer.6.attention.output.dense.weight', 'encoder.layer.6.attention.self.key.bias', 'encoder.layer.6.attention.self.key.weight', 'encoder.layer.6.attention.self.query.bias', 'encoder.layer.6.attention.self.query.weight', 'encoder.layer.6.attention.self.value.bias', 'encoder.layer.6.attention.self.value.weight', 'encoder.layer.6.intermediate.dense.bias', 'encoder.layer.6.intermediate.dense.weight', 'encoder.layer.6.output.LayerNorm.bias', 'encoder.layer.6.output.LayerNorm.weight', 'encoder.layer.6.output.dense.bias', 'encoder.layer.6.output.dense.weight', 'encoder.layer.7.attention.output.LayerNorm.bias', 'encoder.layer.7.attention.output.LayerNorm.weight', 'encoder.layer.7.attention.output.dense.bias', 'encoder.layer.7.attention.output.dense.weight', 'encoder.layer.7.attention.self.key.bias', 'encoder.layer.7.attention.self.key.weight', 'encoder.layer.7.attention.self.query.bias', 'encoder.layer.7.attention.self.query.weight', 'encoder.layer.7.attention.self.value.bias', 'encoder.layer.7.attention.self.value.weight', 'encoder.layer.7.intermediate.dense.bias', 'encoder.layer.7.intermediate.dense.weight', 'encoder.layer.7.output.LayerNorm.bias', 'encoder.layer.7.output.LayerNorm.weight', 'encoder.layer.7.output.dense.bias', 'encoder.layer.7.output.dense.weight', 'encoder.layer.8.attention.output.LayerNorm.bias', 'encoder.layer.8.attention.output.LayerNorm.weight', 'encoder.layer.8.attention.output.dense.bias', 'encoder.layer.8.attention.output.dense.weight', 'encoder.layer.8.attention.self.key.bias', 'encoder.layer.8.attention.self.key.weight', 'encoder.layer.8.attention.self.query.bias', 'encoder.layer.8.attention.self.query.weight', 'encoder.layer.8.attention.self.value.bias', 'encoder.layer.8.attention.self.value.weight', 'encoder.layer.8.intermediate.dense.bias', 'encoder.layer.8.intermediate.dense.weight', 'encoder.layer.8.output.LayerNorm.bias', 'encoder.layer.8.output.LayerNorm.weight', 'encoder.layer.8.output.dense.bias', 'encoder.layer.8.output.dense.weight', 'encoder.layer.9.attention.output.LayerNorm.bias', 'encoder.layer.9.attention.output.LayerNorm.weight', 'encoder.layer.9.attention.output.dense.bias', 'encoder.layer.9.attention.output.dense.weight', 'encoder.layer.9.attention.self.key.bias', 'encoder.layer.9.attention.self.key.weight', 'encoder.layer.9.attention.self.query.bias', 'encoder.layer.9.attention.self.query.weight', 'encoder.layer.9.attention.self.value.bias', 'encoder.layer.9.attention.self.value.weight', 'encoder.layer.9.intermediate.dense.bias', 'encoder.layer.9.intermediate.dense.weight', 'encoder.layer.9.output.LayerNorm.bias', 'encoder.layer.9.output.LayerNorm.weight', 'encoder.layer.9.output.dense.bias', 'encoder.layer.9.output.dense.weight']\n", "You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.\n" ] } ], "source": [ "# Load model directly\n", "from transformers import AutoTokenizer, AutoModelForSequenceClassification\n", "\n", "tokenizer = AutoTokenizer.from_pretrained(\"eunJ/codebert_vulnerability_detector_multi\")\n", "model = AutoModelForSequenceClassification.from_pretrained(\"eunJ/codebert_vulnerability_detector_multi\")" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [] } ], "metadata": { "kernelspec": { "display_name": "xcodevul", "language": "python", "name": "python3" }, "language_info": { "codemirror_mode": { "name": "ipython", "version": 3 }, "file_extension": ".py", "mimetype": "text/x-python", "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", "version": "3.13.5" } }, "nbformat": 4, "nbformat_minor": 2 }