feat: Add sponsorship and website section
Browse files
README.md
CHANGED
@@ -25,6 +25,17 @@ This model builds upon the powerful Qwen2.5-0.5B base model, which features:
|
|
25 |
- 490M parameters (360M non-embedding parameters)
|
26 |
- 24 transformer layers
|
27 |
- 14 attention heads for queries and 2 for key/values (GQA architecture)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
28 |
- Support for 32,768 context length
|
29 |
- Advanced features like RoPE positional embeddings, SwiGLU activations, and RMSNorm
|
30 |
|
|
|
25 |
- 490M parameters (360M non-embedding parameters)
|
26 |
- 24 transformer layers
|
27 |
- 14 attention heads for queries and 2 for key/values (GQA architecture)
|
28 |
+
|
29 |
+
---
|
30 |
+
### 🌐 Website
|
31 |
+
You can find more of my models, projects, and information on my official website:
|
32 |
+
- **[artificialguy.com](https://artificialguy.com/)**
|
33 |
+
|
34 |
+
### 💖 Support My Work
|
35 |
+
If you find this model useful, please consider supporting my work. It helps me cover server costs and dedicate more time to new open-source projects.
|
36 |
+
- **Patreon:** [Support on Patreon](https://www.patreon.com/user?u=81570187)
|
37 |
+
- **Ko-fi:** [Buy me a Ko-fi](https://ko-fi.com/artificialguybr)
|
38 |
+
- **Buy Me a Coffee:** [Buy me a Coffee](https://buymeacoffee.com/jvkape)
|
39 |
- Support for 32,768 context length
|
40 |
- Advanced features like RoPE positional embeddings, SwiGLU activations, and RMSNorm
|
41 |
|