--- license: cc-by-nc-4.0 pipeline_tag: text-generation tags: - moe --- ![img_text](./assets/tmpd2xdo_x4.png) Proctora is a MoE model made of - OpenPipe/mistral-ft-optimized-1227 as a base model - SanjiWatsuki/Kunoichi-7B as a first expert dedicated to RP tasks. - samir-fama/SamirGPT-v1 as a second expert for factual answers. I do not have yet any metrics for this model but subjective ones. It was made at first out of curiosity and experimentation. My goal is to produce a model excellent at being a game master for RPG sessions. However being dissatisfied with the existing evaluation tool-sets, I decided to create my own (still a WIP on 01/16/24). And among my collection of models of small/medium models Proctora gave me the best results to evaluate the answers produced by other LLMs. Therefore, I surprisingly settled with it and gave it an appropriate name according to the task. TLDR: Proctora is a tool for a tool! I doubt this model will be useful for the community. I publish it for the sake of transparency in my creative process.