YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
MoE model based on Ling V2 MoE arch, with 32 experts, 2 activated, no shared experts, no dense layers, trained on about 100M tokens of FineWeb-2 pol-Latn split, using tokenizer taken from EuroLLM 1.7B
Just a test to validate pipeline before training bigger 4B 256 expert model on 100B tokens.
- Downloads last month
- 41
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support