zhiyuan8 commited on
Commit
004140b
·
verified ·
1 Parent(s): bbda20d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -65,11 +65,9 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
65
  model = AutoModelForCausalLM.from_pretrained('fla-hub/rwkv7-1.5B-world', trust_remote_code=True)
66
  tokenizer = AutoTokenizer.from_pretrained('fla-hub/rwkv7-1.5B-world', trust_remote_code=True)
67
 
68
- model = model.cuda()
69
  prompt = "What is a large language model?"
70
  messages = [
71
- {"role": "user", "content": "Who are you?"},
72
- {"role": "assistant", "content": "I am a GPT-3 based model."},
73
  {"role": "user", "content": prompt}
74
  ]
75
  text = tokenizer.apply_chat_template(
@@ -82,7 +80,11 @@ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
82
 
83
  generated_ids = model.generate(
84
  **model_inputs,
85
- max_new_tokens=1024,
 
 
 
 
86
  )
87
  generated_ids = [
88
  output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
 
65
  model = AutoModelForCausalLM.from_pretrained('fla-hub/rwkv7-1.5B-world', trust_remote_code=True)
66
  tokenizer = AutoTokenizer.from_pretrained('fla-hub/rwkv7-1.5B-world', trust_remote_code=True)
67
 
68
+ model = model.cuda() # Supported on Nvidia/AMD/Intel eg. model.xpu()
69
  prompt = "What is a large language model?"
70
  messages = [
 
 
71
  {"role": "user", "content": prompt}
72
  ]
73
  text = tokenizer.apply_chat_template(
 
80
 
81
  generated_ids = model.generate(
82
  **model_inputs,
83
+ max_new_tokens=4096,
84
+ do_sample=True,
85
+ temperature=1.0,
86
+ top_p=0.3,
87
+ repetition_penalty=1.2
88
  )
89
  generated_ids = [
90
  output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)