Lowenzahn commited on
Commit
5997d1b
·
verified ·
1 Parent(s): 62edf9f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -16
README.md CHANGED
@@ -5,12 +5,18 @@ language:
5
  - en
6
  base_model:
7
  - meta-llama/Llama-3.1-8B
 
 
 
 
8
  ---
9
 
10
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/646704281dd5854d4de2cdda/vBbfKhWIBJVIjqSv6yVDk.png)
11
 
12
  # KoBioMed-Llama-3.1-8B
13
 
 
 
14
  ## Instroduction
15
  We introduce KoBioMed-Llama-3.1-8B, a bilingual (English and Korean) generative model specialized in the BioMedical domain, developed by ezCaretech. This model has been continual pre-trained (CPT) on a dataset from PubMed abstracts and their translated Korean counterparts, undergoing extensive preprocessing that includes cleansing, de-duplication, and quality filtering.
16
 
@@ -24,12 +30,16 @@ This repository contains an 8 Billion generative language model with the followi
24
  - Vocab Size: 12,800
25
  - License: llama3.1
26
 
27
- **This model was developed with support from the Korea Artificial Intelligence Industry Cluster Agency (AICA).**
28
 
29
 
30
 
31
  ### Notice!
32
- The model is currently in post-training (Instruction Tuning, DPO) and is scheduled to be released within March 2025.
 
 
 
 
33
 
34
 
35
  ## Evaluation
@@ -48,23 +58,23 @@ We evaluated the KoBioMed-Llama-3.1-8B using various Korean and English biomedic
48
  <th>MedMCQA</th>
49
  <th>MMLU</th>
50
  <th>PubMedQA</th>
51
- <th>Average</th>
52
  </tr>
53
  <tr>
54
- <td>KoBioMed-Llama-3.1-8B</td>
55
- <td align="center"><strong>0.0</strong></td>
56
- <td align="center"><strong>0.0</strong></td>
57
- <td align="center"><strong>0.0</strong></td>
58
- <td align="center"><strong>0.0</strong></td>
59
- <td align="center"><strong>0.0</strong></td>
60
- <td align="center"><strong>0.0</strong></td>
61
  </tr>
62
  <tr>
63
  <td>Llama-3.1-8B</td>
64
- <td align="center">0.3620</td>
65
  <td align="center">0.5105</td>
66
- <td align="center">0.5635</td>
67
- <td align="center">0.7159</td>
68
  <td align="center">0.7600</td>
69
  <td align="center">0.5824</td>
70
  </tr>
@@ -89,10 +99,10 @@ We evaluated the KoBioMed-Llama-3.1-8B using various Korean and English biomedic
89
  <tr>
90
  <td>SOLAR-10.7B-v1.0</td>
91
  <td align="center">0.3200</td>
92
- <td align="center">0.5146</td>
93
  <td align="center">0.5075</td>
94
- <td align="center">0.7050</td>
95
- <td align="center">0.7760</td>
96
  <td align="center">0.5646</td>
97
  </tr>
98
  </table>
 
5
  - en
6
  base_model:
7
  - meta-llama/Llama-3.1-8B
8
+ tags:
9
+ - biology
10
+ - medical
11
+ - Llama
12
  ---
13
 
14
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/646704281dd5854d4de2cdda/vBbfKhWIBJVIjqSv6yVDk.png)
15
 
16
  # KoBioMed-Llama-3.1-8B
17
 
18
+ _**The model weights are scheduled to be uploaded on March 17, 2025.**_
19
+
20
  ## Instroduction
21
  We introduce KoBioMed-Llama-3.1-8B, a bilingual (English and Korean) generative model specialized in the BioMedical domain, developed by ezCaretech. This model has been continual pre-trained (CPT) on a dataset from PubMed abstracts and their translated Korean counterparts, undergoing extensive preprocessing that includes cleansing, de-duplication, and quality filtering.
22
 
 
30
  - Vocab Size: 12,800
31
  - License: llama3.1
32
 
33
+
34
 
35
 
36
 
37
  ### Notice!
38
+ - **This is a pre-trained model. It will be a great starting point for post-training, such as instruction tuning.**
39
+
40
+ - **This model was developed with support from the Korea Artificial Intelligence Industry Cluster Agency (AICA).**
41
+
42
+ - **The model is currently in post-training (Instruction Tuning, DPO) and is scheduled to be released within March 2025.**
43
 
44
 
45
  ## Evaluation
 
58
  <th>MedMCQA</th>
59
  <th>MMLU</th>
60
  <th>PubMedQA</th>
61
+ <th>Mean</th>
62
  </tr>
63
  <tr>
64
+ <td>KoBioMed-Llama-3.1-8B (v0.5)</td>
65
+ <td align="center"><strong>0.3930</strong></td>
66
+ <td align="center"><strong>0.5641</strong></td>
67
+ <td align="center"><u>0.5326</u></td>
68
+ <td align="center">0.6771</td>
69
+ <td align="center"><strong>0.7940</strong></td>
70
+ <td align="center"><strong>0.5922</strong></td>
71
  </tr>
72
  <tr>
73
  <td>Llama-3.1-8B</td>
74
+ <td align="center"><u>0.3620</u></td>
75
  <td align="center">0.5105</td>
76
+ <td align="center"><strong>0.5635</strong></td>
77
+ <td align="center"><strong>0.7159</strong></td>
78
  <td align="center">0.7600</td>
79
  <td align="center">0.5824</td>
80
  </tr>
 
99
  <tr>
100
  <td>SOLAR-10.7B-v1.0</td>
101
  <td align="center">0.3200</td>
102
+ <td align="center"><u>0.5146</u></td>
103
  <td align="center">0.5075</td>
104
+ <td align="center"><u>0.7050</u></td>
105
+ <td align="center"><u>0.7760</u></td>
106
  <td align="center">0.5646</td>
107
  </tr>
108
  </table>