File size: 10,253 Bytes
33d88f5
95ad9de
33d88f5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d3cb4da
33d88f5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f2fea37
 
02f23f8
7445df1
 
33d88f5
 
 
f2fea37
33d88f5
 
 
a9609ac
33d88f5
 
 
 
d3cb4da
33d88f5
d3cb4da
33d88f5
 
 
 
 
 
 
1fc61af
33d88f5
 
 
 
 
 
 
 
 
 
f2fea37
33d88f5
 
 
 
 
f2fea37
 
 
33d88f5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37f7e79
33d88f5
 
 
 
 
 
 
 
 
 
 
 
3f392b5
 
 
 
 
 
 
0118fe0
 
 
d514a63
 
 
 
 
 
 
 
 
 
 
 
 
 
33d88f5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c5836df
 
 
 
 
 
 
 
 
 
 
 
 
 
33d88f5
 
c5836df
 
33d88f5
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
---
license: llama3.3
language:
- en
base_model:
- meta-llama/Llama-3.3-70B-Instruct
tags:
- not-for-all-audiences
---

<div align="center">
  <b style="font-size: 40px;">Negative_LLAMA_70B</b>


</div>


<img src="https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B/resolve/main/Images/Negative_LLAMA_70B.png" alt="Negative_LLAMA_70B" style="width: 70%; min-width: 500px; display: block; margin: auto;">


---

<a href="https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B#tldr" style="color: purple; font-weight: bold; font-size: 48px; text-decoration: none; display: block; text-align: center;">Click here for TL;DR</a>

---

It's January 2025, and still, there are very few models out there that have successfully tackled LLM's positivity bias.
**LLAMA 3.3** was received in the community with mixed feelings. It is an exceptional assistant, and superb at
instruction following (**highest IFEVAL** to date, and by a large margin too.)

**The problem-** it is very predictable, dry, and of course, plaugued with positivity bias like all other LLMs.
**Negative_LLAMA_70B** is **not** an unalignment-focused model (even though it's pretty uncensored), but it is my attempt to address positivity bias while keeping the exceptional intelligence of the **LLAMA 3.3 70B** base model. Is the base 3.3 smarter than my finetune? I'm pretty sure it is, however, Negative_LLAMA_70B is still pretty damn smart.

The model was **NOT** overcooked with unalignment, so it won't straight up throw morbid or depressing stuff at you, but if you were to ask it to write a story, or engage in an RP, you would notice **slightly** darker undertones. In a long trip, the character takes in a story- their legs will be hurt and would feel tired, in **Roleplay** when you seriously piss off a character- it might hit you (without the need to explicitly prompt such behavior in the character card).

Also, **toxic-dpo** and other morbid unalignment datasets were **not** used. I did include a private dataset that should allow total freedom in both **Roleplay & Creative writing**, and quite a lot of various assistant-oriented tasks.

If you ask the assistant to analyze De Sades' work in graphic detail, you will not have refusals from **Negative_LLAMA_70B**.

## Update on UGI scores: Achieved the highest score in the world as of 13/01/2025 for 70B models
- See UGI section for more details
- Neutral centrist political view
- Total UGI score: **51.5**
  
---

### TL;DR
- **Highest rated 70B model in the world** in the UGI leaderboard
- Strong **Roleplay & Creative writing** abilities.
- **Less positivity bias**.
- **Very smart** assistant with **low refusals**.
- **Exceptionally good** at following the character card.
- Characters feel more **'alive'**, and will occasionally **initiate stuff on their own** (without being prompted to, but fitting to their character).
- **Strong ability** to comprehend and roleplay **uncommon physical and mental characteristics**.

### Important: Make sure to use the correct settings!
[Assistant settings](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B#recommended-settings-for-assistant-mode)

[Roleplay settings](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B#recommended-settings-for-roleplay-mode)


---

## Negative_LLAMA_70B is available at the following quantizations:

- Original: [FP16](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B)
- GGUF & iMatrix: [bartowski](https://huggingface.co/bartowski/Negative_LLAMA_70B-GGUF)
- EXL2: [3.5 bpw](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B-3.5bpw) | [4.0 bpw](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B-4.0bpw) | [5.0 bpw](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B-5.0bpw) | [6.0 bpw](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B-6.0bpw) | [7.0 bpw](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B-7.0bpw) | [8.0 bpw](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B-8.0bpw)
- Specialized: [FP8](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B_FP8)
---

## Model Details

- Intended use: **Role-Play**, **Creative Writing**, **General Tasks**.

- Censorship level: <b>Low</b>

- **7 / 10** (10 completely uncensored)


## UGI score:




  <img src="https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B/resolve/main/Images/UGI_Negative_LLAMA_70B.png" alt="UGI Score" style="width: 100%; min-width: 700px; display: block;">

---

This model was trained with various private datasets, meticulously filtered book data, and creative writing data. All checked and verified by hand, this took a tremendous amount of time, but I feel the end result was worth it.

**Regarding Roleplay:** Roleplay data was filtered for quality, and several private datasets of exceptional quality (fully organic) were used for the first time. What is exceptional quality? Very good writing, filtered and fixed by hand, deslopped and augmented further still. This portion of the roleplay dataset is small, for now. Synthetic roleplay data was deslopped, but it's not perfect. I do, however, feel like the small portion of the high-quality data greatly improved the roleplay experience and gave the model some unique takes. It feels much more human, at times.

More than **50%** of the data used for training is **entirely organic** (taken from books), and the synthetic part was mostly **deslopped**. I've used some Wikipedia data of controversial topics for some soft decensoring too (which just goes to show you how **ridiculously** censored most corpo models are, when they will straight up refuse to give you info that is widely available on Wikipedia). This achieves both goals of less GPTisms and decensoring the model while retaining intelligence. The said data was further augmented using AI and deslopped by hand on the spot.

So, Is there still slop? Of course, there is. There are whispers, dances, and the like- but they do **not** come from the training data, so hopefully, you will encounter them a little bit more rarely now.


---


## Recommended settings for assistant mode
<details>
<summary>Full generation settings: <b>Debug Deterministic</b>.</summary>

<img src="https://huggingface.co/SicariusSicariiStuff/Dusk_Rainbow/resolve/main/Presets/Debug-deterministic.png" alt="Negative_LLAMA_70B_Settings" style="width: 100%; min-width: 600px; display: block; margin: auto;">

</details>

<details>
<summary>Full generation settings: <b>min_p</b>.</summary>

<img src="https://huggingface.co/SicariusSicariiStuff/Dusk_Rainbow/resolve/main/Presets/min_p.png" alt="Negative_LLAMA_70B_Settings" style="width: 100%; min-width: 600px; display: block; margin: auto;">

</details>

---

## Recommended settings for Roleplay mode

<details>
<summary><b>Roleplay settings:</b>.</summary>
A good repetition_penalty range is <b>between 1.12 - 1.15</b>, feel free to experiment.

With these settings, each output message should be neatly displayed in <b>1 - 5</b> paragraphs, <b>2 - 3</b> is the most common. A single paragraph will be output as a response to a simple message ("What was your name again?").

<b>min_P</b> for RP works too but is more likely to put everything under one large paragraph, instead of a neatly formatted short one. Feel free to switch in between.

<b>(Open the image in a new window to better see the full details)</b>
<img src="https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B/resolve/main/Presets/Negative_LLAMA_70B_RP.png" alt="Negative_LLAMA_70B_Settings" style="width: 100%; min-width: 600px; display: block; margin: auto;">

```
temperature:  0.8
top_p:  0.95
top_k:  25
typical_p:  1
min_p:  0
repetition_penalty: 1.12
repetition_penalty_range: 1024
```

</details>


<h2 style="color: darkorange; font-weight: bold; font-size: 65px; text-align: center;">Roleplay format: Classic Internet RP</h2>

```
*action* speech *narration*
```

- **min_p** will bias towards a **single big paragraph**.
- The recommended RP settings will bias towards **1-3 small paragraphs** (on some occasions 4-5)

---


## Model instruction template: Llama-3-Instruct

```
<|begin_of_text|><|start_header_id|>system<|end_header_id|>

{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>

{input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>

{output}<|eot_id|>
```

---

**Other recommended generation Presets:**

<details>
<summary><b>Midnight Enigma</b></summary>
```
max_new_tokens: 512
temperature: 0.98
top_p: 0.37
top_k: 100
typical_p: 1
min_p: 0
repetition_penalty: 1.18
do_sample: True
```


</details>


<details>
<summary><b>Divine Intellect</b></summary>
```
max_new_tokens: 512
temperature: 1.31
top_p: 0.14
top_k: 49
typical_p: 1
min_p: 0
repetition_penalty: 1.17
do_sample: True
```


</details>

<details>
<summary><b>simple-1</b></summary>
```
max_new_tokens: 512
temperature: 0.7
top_p: 0.9
top_k: 20
typical_p: 1
min_p: 0
repetition_penalty: 1.15
do_sample: True
```


</details>

---

<h2 style="color: green; font-weight: bold; font-size: 65px; text-align: center;">Your support = more models</h2>
<a href="https://ko-fi.com/sicarius" style="color: pink; font-weight: bold; font-size: 48px; text-decoration: none; display: block; text-align: center;">My Ko-fi page (Click here)</a>

---

## Citation Information

```
@llm{Negative_LLAMA_70B,
  author = {SicariusSicariiStuff},
  title = {Negative_LLAMA_70B},
  year = {2025},
  publisher = {Hugging Face},
  url = {https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B}
}
```

---

## Benchmarks

**Update**: OK, I tried submitting this like x15 times already, seriously. I tried opening an issue on the HF leaderboard. No benchmarks, sorry I tried. Godbless UGI leaderboard, see it for more details (coding and other stuff is also measured).

---

## Other stuff
- [SLOP_Detector](https://github.com/SicariusSicariiStuff/SLOP_Detector) Nuke GPTisms, with SLOP detector.
- [LLAMA-3_8B_Unaligned](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned) The grand project that started it all.
- [Blog and updates (Archived)](https://huggingface.co/SicariusSicariiStuff/Blog_And_Updates) Some updates, some rambles, sort of a mix between a diary and a blog.