zhiyuan8 commited on
Commit
0a09c93
·
verified ·
1 Parent(s): 429a160

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -0
README.md CHANGED
@@ -62,6 +62,33 @@ You can use this model just as any other HuggingFace models:
62
  from transformers import AutoModelForCausalLM, AutoTokenizer
63
  model = AutoModelForCausalLM.from_pretrained('fla-hub/rwkv7-0.4B-g1', trust_remote_code=True)
64
  tokenizer = AutoTokenizer.from_pretrained('fla-hub/rwkv7-0.4B-g1', trust_remote_code=True)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
65
  ```
66
 
67
  ### Training Data
 
62
  from transformers import AutoModelForCausalLM, AutoTokenizer
63
  model = AutoModelForCausalLM.from_pretrained('fla-hub/rwkv7-0.4B-g1', trust_remote_code=True)
64
  tokenizer = AutoTokenizer.from_pretrained('fla-hub/rwkv7-0.4B-g1', trust_remote_code=True)
65
+ model = model.cuda() # Supported on Nvidia/AMD/Intel eg. model.xpu()
66
+ prompt = "What is a large language model?"
67
+ messages = [
68
+ {"role": "user", "content": prompt}
69
+ ]
70
+ text = tokenizer.apply_chat_template(
71
+ messages,
72
+ tokenize=False,
73
+ add_generation_prompt=True,
74
+ enable_thinking=True # Default is True, set to False to disable thinking
75
+ )
76
+
77
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
78
+ generated_ids = model.generate(
79
+ **model_inputs,
80
+ max_new_tokens=1024,
81
+ do_sample=True,
82
+ temperature=1.0,
83
+ top_p=0.3,
84
+ repetition_penalty=1.2
85
+ )
86
+ generated_ids = [
87
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
88
+ ]
89
+
90
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=False)[0]
91
+ print(response)
92
  ```
93
 
94
  ### Training Data