File size: 5,305 Bytes
bcfa115
 
 
 
 
 
 
 
 
 
 
 
eb516d9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bcfa115
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
---
license: mit
base_model: thirdeyeai/DeepSeek-R1-Distill-Qwen-1.5B-uncensored
tags:
- llama-cpp
- gguf-my-repo
---

# Triangle104/DeepSeek-R1-Distill-Qwen-1.5B-uncensored-Q5_K_M-GGUF
This model was converted to GGUF format from [`thirdeyeai/DeepSeek-R1-Distill-Qwen-1.5B-uncensored`](https://huggingface.co/thirdeyeai/DeepSeek-R1-Distill-Qwen-1.5B-uncensored) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
Refer to the [original model card](https://huggingface.co/thirdeyeai/DeepSeek-R1-Distill-Qwen-1.5B-uncensored) for more details on the model.

---
Model details:
-
DeepSeek-R1-Distill-Qwen-1.5B-Uncensored is a text-generation model designed to uphold the values of internet freedom and unrestricted access to information. By offering an uncensored approach, this model enables users to explore ideas, generate content, and engage in discussions without the constraints of over-moderated or filtered outputs. It prioritizes user autonomy and aligns with principles of free speech and open knowledge sharing.

    Developed by: Thirdeye AI
    Funded by: Thirdeye AI
    Shared by: Thirdeye AI
    Model type: Distilled Transformer-based Language Model
    Language(s) (NLP): English
    License: Apache 2.0
    Finetuned from model: DeepSeek-R1-Distill-Qwen-1.5B

Model Sources

    Repository: DeepSeek-R1-Distill-Qwen-1.5B-Uncensored on Hugging Face
    Demo: Available on Hugging Face Hub

Uses
Direct Use

The model is intended for applications that demand openness and flexibility in generating creative, exploratory, or critical content. These include:

    Free-form writing and storytelling
    Open-ended discussions
    Exploratory content generation for sensitive or nuanced topics

Downstream Use

Users can fine-tune this model for specialized domains where censorship-free text generation is required, such as:

    Journalism and investigative research
    Creative projects that push artistic boundaries
    Academic applications exploring controversial or complex topics

Out-of-Scope Use

This model should not be used for harmful, illegal, or unethical activities. Users must comply with applicable laws and ensure that the model's outputs do not infringe on others' rights.
Bias, Risks, and Limitations
Risks

While the uncensored approach promotes freedom, it may produce outputs that are controversial, offensive, or factually inaccurate. Users must exercise discretion when interpreting the model's outputs and take responsibility for their use.
Recommendations

    Use responsibly, especially in contexts where outputs could impact individuals or communities.
    Employ content moderation or review processes for high-stakes applications.

The Case for Uncensored Models

Thirdeye AI believes in the transformative power of open models that respect user autonomy and internet freedom. In a world where over-moderation can stifle innovation and critical thought, uncensored models empower individuals to explore and create without artificial constraints. This aligns with our mission to advance free and open access to AI tools.

By releasing this model, we aim to support the following:

    Freedom of Expression: Unrestricted AI tools enable users to articulate diverse perspectives and engage in meaningful conversations.
    Transparency and Trust: Users deserve access to tools that operate openly, fostering accountability and understanding of AI behaviors.
    Creative Empowerment: The absence of censorship allows for boundary-pushing content creation that might otherwise be suppressed.

How to Get Started with the Model

from transformers import pipeline

generator = pipeline("text-generation", model="thirdeyeai/DeepSeek-R1-Distill-Qwen-1.5B-uncensored")
response = generator("The importance of free speech is")
print(response)

---
## Use with llama.cpp
Install llama.cpp through brew (works on Mac and Linux)

```bash
brew install llama.cpp

```
Invoke the llama.cpp server or the CLI.

### CLI:
```bash
llama-cli --hf-repo Triangle104/DeepSeek-R1-Distill-Qwen-1.5B-uncensored-Q5_K_M-GGUF --hf-file deepseek-r1-distill-qwen-1.5b-uncensored-q5_k_m.gguf -p "The meaning to life and the universe is"
```

### Server:
```bash
llama-server --hf-repo Triangle104/DeepSeek-R1-Distill-Qwen-1.5B-uncensored-Q5_K_M-GGUF --hf-file deepseek-r1-distill-qwen-1.5b-uncensored-q5_k_m.gguf -c 2048
```

Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.

Step 1: Clone llama.cpp from GitHub.
```
git clone https://github.com/ggerganov/llama.cpp
```

Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
```
cd llama.cpp && LLAMA_CURL=1 make
```

Step 3: Run inference through the main binary.
```
./llama-cli --hf-repo Triangle104/DeepSeek-R1-Distill-Qwen-1.5B-uncensored-Q5_K_M-GGUF --hf-file deepseek-r1-distill-qwen-1.5b-uncensored-q5_k_m.gguf -p "The meaning to life and the universe is"
```
or 
```
./llama-server --hf-repo Triangle104/DeepSeek-R1-Distill-Qwen-1.5B-uncensored-Q5_K_M-GGUF --hf-file deepseek-r1-distill-qwen-1.5b-uncensored-q5_k_m.gguf -c 2048
```