{ "nbformat": 4, "nbformat_minor": 0, "metadata": { "colab": { "private_outputs": true, "provenance": [], "gpuType": "T4", "authorship_tag": "ABX9TyOv14c2MWENhO6RJ3uy6vD7", "include_colab_link": true }, "kernelspec": { "name": "python3", "display_name": "Python 3" }, "language_info": { "name": "python" }, "accelerator": "GPU" }, "cells": [ { "cell_type": "markdown", "metadata": { "id": "view-in-github", "colab_type": "text" }, "source": [ "\"Open" ] }, { "cell_type": "code", "execution_count": null, "metadata": { "cellView": "form", "id": "uJS9i_Dltv8Y" }, "outputs": [], "source": [ "#@title v-- Enter your model below and then click this to start Koboldcpp\n", "\n", "Model = \"https://huggingface.co/TheBloke/airoboros-l2-13B-gpt4-1.4.1-GGML/resolve/main/airoboros-l2-13b-gpt4-1.4.1.ggmlv3.q4_0.bin\" #@param [\"\"]{allow-input: true}\n", "Layers = 43 #@param [43]{allow-input: true}\n", "\n", "%cd /content\n", "!git clone https://github.com/LostRuins/koboldcpp\n", "%cd /content/koboldcpp\n", "!make LLAMA_CUBLAS=1\n", "\n", "!wget $Model -O model.ggml\n", "!wget -c https://github.com/cloudflare/cloudflared/releases/latest/download/cloudflared-linux-amd64\n", "!chmod +x cloudflared-linux-amd64\n", "!nohup ./cloudflared-linux-amd64 tunnel --url http://localhost:5001 &\n", "!sleep 10\n", "!cat nohup.out\n", "!python koboldcpp.py model.ggml --stream --usecublas 0 --gpulayers $Layers --hordeconfig concedo\n" ] } ] }