Ramikan-BR's picture
Update README.md
6aa01b9 verified
|
raw
history blame
2.42 kB
metadata
language:
  - en
license: apache-2.0
tags:
  - text-generation-inference
  - transformers
  - unsloth
  - llama
  - trl
  - sft
  - code
  - lora
  - peft
base_model: unsloth/tinyllama-chat-bnb-4bit
pipeline_tag: text-generation
datasets: Ramikan-BR/data-oss_instruct-decontaminated_python.jsonl

Uploaded model

  • Developed by: Ramikan-BR
  • Model type: [text-generation/Python Coder]
  • Language(s) (NLP): [en]
  • License: apache-2.0
  • Finetuned from model : unsloth/tinyllama-chat-bnb-4bit

Model Description

Training Data

datasets: Ramikan-BR/data-oss_instruct-decontaminated_python.jsonl

Training Procedure

The model was refined using Unsloath. The dataset ise-uiuc/Magicoder-OSS-Instruct-75K was adjusted, leaving only data on python and divided into 10 parts, each refinement occurred for 2 epochs, using adafactor optimizer or adamw_8bit (adafactor seems to deliver less loss).

Model Sources [optional]

base_model: unsloth/tinyllama-chat-bnb-4bit

model: Ramikan-BR/tinyllama-coder-py-4bit-v10 gguf_f16: tinyllama-coder-py-4bit-v10-unsloth.F16.gguf gguf_Q4_K_M: tinyllama-coder-py-4bit-v10-unsloth.Q4_K_M.gguf gguf_Q8_0: tinyllama-coder-py-4bit-v10-unsloth.Q8_0.gguf

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.