Update README.md
Browse files
README.md
CHANGED
@@ -5,12 +5,18 @@ language:
|
|
5 |
- en
|
6 |
base_model:
|
7 |
- meta-llama/Llama-3.1-8B
|
|
|
|
|
|
|
|
|
8 |
---
|
9 |
|
10 |

|
11 |
|
12 |
# KoBioMed-Llama-3.1-8B
|
13 |
|
|
|
|
|
14 |
## Instroduction
|
15 |
We introduce KoBioMed-Llama-3.1-8B, a bilingual (English and Korean) generative model specialized in the BioMedical domain, developed by ezCaretech. This model has been continual pre-trained (CPT) on a dataset from PubMed abstracts and their translated Korean counterparts, undergoing extensive preprocessing that includes cleansing, de-duplication, and quality filtering.
|
16 |
|
@@ -24,12 +30,16 @@ This repository contains an 8 Billion generative language model with the followi
|
|
24 |
- Vocab Size: 12,800
|
25 |
- License: llama3.1
|
26 |
|
27 |
-
|
28 |
|
29 |
|
30 |
|
31 |
### Notice!
|
32 |
-
|
|
|
|
|
|
|
|
|
33 |
|
34 |
|
35 |
## Evaluation
|
@@ -48,23 +58,23 @@ We evaluated the KoBioMed-Llama-3.1-8B using various Korean and English biomedic
|
|
48 |
<th>MedMCQA</th>
|
49 |
<th>MMLU</th>
|
50 |
<th>PubMedQA</th>
|
51 |
-
<th>
|
52 |
</tr>
|
53 |
<tr>
|
54 |
-
<td>KoBioMed-Llama-3.1-8B</td>
|
55 |
-
<td align="center"><strong>0.
|
56 |
-
<td align="center"><strong>0.
|
57 |
-
<td align="center"><
|
58 |
-
<td align="center"
|
59 |
-
<td align="center"><strong>0.
|
60 |
-
<td align="center"><strong>0.
|
61 |
</tr>
|
62 |
<tr>
|
63 |
<td>Llama-3.1-8B</td>
|
64 |
-
<td align="center">0.3620</td>
|
65 |
<td align="center">0.5105</td>
|
66 |
-
<td align="center">0.5635</td>
|
67 |
-
<td align="center">0.7159</td>
|
68 |
<td align="center">0.7600</td>
|
69 |
<td align="center">0.5824</td>
|
70 |
</tr>
|
@@ -89,10 +99,10 @@ We evaluated the KoBioMed-Llama-3.1-8B using various Korean and English biomedic
|
|
89 |
<tr>
|
90 |
<td>SOLAR-10.7B-v1.0</td>
|
91 |
<td align="center">0.3200</td>
|
92 |
-
<td align="center">0.5146</td>
|
93 |
<td align="center">0.5075</td>
|
94 |
-
<td align="center">0.7050</td>
|
95 |
-
<td align="center">0.7760</td>
|
96 |
<td align="center">0.5646</td>
|
97 |
</tr>
|
98 |
</table>
|
|
|
5 |
- en
|
6 |
base_model:
|
7 |
- meta-llama/Llama-3.1-8B
|
8 |
+
tags:
|
9 |
+
- biology
|
10 |
+
- medical
|
11 |
+
- Llama
|
12 |
---
|
13 |
|
14 |

|
15 |
|
16 |
# KoBioMed-Llama-3.1-8B
|
17 |
|
18 |
+
_**The model weights are scheduled to be uploaded on March 17, 2025.**_
|
19 |
+
|
20 |
## Instroduction
|
21 |
We introduce KoBioMed-Llama-3.1-8B, a bilingual (English and Korean) generative model specialized in the BioMedical domain, developed by ezCaretech. This model has been continual pre-trained (CPT) on a dataset from PubMed abstracts and their translated Korean counterparts, undergoing extensive preprocessing that includes cleansing, de-duplication, and quality filtering.
|
22 |
|
|
|
30 |
- Vocab Size: 12,800
|
31 |
- License: llama3.1
|
32 |
|
33 |
+
|
34 |
|
35 |
|
36 |
|
37 |
### Notice!
|
38 |
+
- **This is a pre-trained model. It will be a great starting point for post-training, such as instruction tuning.**
|
39 |
+
|
40 |
+
- **This model was developed with support from the Korea Artificial Intelligence Industry Cluster Agency (AICA).**
|
41 |
+
|
42 |
+
- **The model is currently in post-training (Instruction Tuning, DPO) and is scheduled to be released within March 2025.**
|
43 |
|
44 |
|
45 |
## Evaluation
|
|
|
58 |
<th>MedMCQA</th>
|
59 |
<th>MMLU</th>
|
60 |
<th>PubMedQA</th>
|
61 |
+
<th>Mean</th>
|
62 |
</tr>
|
63 |
<tr>
|
64 |
+
<td>KoBioMed-Llama-3.1-8B (v0.5)</td>
|
65 |
+
<td align="center"><strong>0.3930</strong></td>
|
66 |
+
<td align="center"><strong>0.5641</strong></td>
|
67 |
+
<td align="center"><u>0.5326</u></td>
|
68 |
+
<td align="center">0.6771</td>
|
69 |
+
<td align="center"><strong>0.7940</strong></td>
|
70 |
+
<td align="center"><strong>0.5922</strong></td>
|
71 |
</tr>
|
72 |
<tr>
|
73 |
<td>Llama-3.1-8B</td>
|
74 |
+
<td align="center"><u>0.3620</u></td>
|
75 |
<td align="center">0.5105</td>
|
76 |
+
<td align="center"><strong>0.5635</strong></td>
|
77 |
+
<td align="center"><strong>0.7159</strong></td>
|
78 |
<td align="center">0.7600</td>
|
79 |
<td align="center">0.5824</td>
|
80 |
</tr>
|
|
|
99 |
<tr>
|
100 |
<td>SOLAR-10.7B-v1.0</td>
|
101 |
<td align="center">0.3200</td>
|
102 |
+
<td align="center"><u>0.5146</u></td>
|
103 |
<td align="center">0.5075</td>
|
104 |
+
<td align="center"><u>0.7050</u></td>
|
105 |
+
<td align="center"><u>0.7760</u></td>
|
106 |
<td align="center">0.5646</td>
|
107 |
</tr>
|
108 |
</table>
|