Update README.md
Browse files
README.md
CHANGED
@@ -20,7 +20,8 @@ tags:
|
|
20 |
## Model Overview
|
21 |
Huihui-MoE-23B-A4B-abliterated is a **Mixture of Experts (MoE)** language model developed by **huihui.ai**, built upon the **[huihui-ai/Huihui-Qwen3-4B-abliterated-v2](https://huggingface.co/huihui-ai/Huihui-Qwen3-4B-abliterated-v2)** base model. It enhances the standard Transformer architecture by replacing MLP layers with MoE layers, each containing 8 experts, to achieve high performance with efficient inference. The model is designed for natural language processing tasks, including text generation, question answering, and conversational applications.
|
22 |
|
23 |
-
The corresponding version is [huihui-ai/Huihui-MoE-23B-A4B](https://huggingface.co/huihui-ai/Huihui-MoE-23B-A4B)
|
|
|
24 |
|
25 |
**Note**:
|
26 |
The activated expert can handle numbers from 1 to 8, and can complete normal conversations as well.
|
|
|
20 |
## Model Overview
|
21 |
Huihui-MoE-23B-A4B-abliterated is a **Mixture of Experts (MoE)** language model developed by **huihui.ai**, built upon the **[huihui-ai/Huihui-Qwen3-4B-abliterated-v2](https://huggingface.co/huihui-ai/Huihui-Qwen3-4B-abliterated-v2)** base model. It enhances the standard Transformer architecture by replacing MLP layers with MoE layers, each containing 8 experts, to achieve high performance with efficient inference. The model is designed for natural language processing tasks, including text generation, question answering, and conversational applications.
|
22 |
|
23 |
+
The corresponding version is [huihui-ai/Huihui-MoE-23B-A4B](https://huggingface.co/huihui-ai/Huihui-MoE-23B-A4B).
|
24 |
+
Here, simply by changing the instruction version to the ablation version, we can achieve the ablation effect. This is just a try for testing, and we can compare the differences between these two versions.
|
25 |
|
26 |
**Note**:
|
27 |
The activated expert can handle numbers from 1 to 8, and can complete normal conversations as well.
|