Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ language:
|
|
5 |
base_model:
|
6 |
- Lin-Chen/open-llava-next-llama3-8b
|
7 |
tags:
|
8 |
-
- remote
|
9 |
---
|
10 |
# Adapting Multimodal Large Language Models to Domains via Post-Training
|
11 |
|
@@ -28,7 +28,7 @@ image = Image.open(requests.get(url, stream=True).raw).convert("RGB")
|
|
28 |
|
29 |
instruction = "What's in the image?"
|
30 |
|
31 |
-
model_path='AdaptLLM/remote
|
32 |
|
33 |
# =========================== Do NOT need to modify the following ===============================
|
34 |
# Load the processor
|
@@ -60,7 +60,7 @@ print(pred)
|
|
60 |
|
61 |
## 2. To Evaluate Any MLLM on Domain-Specific Benchmarks
|
62 |
|
63 |
-
Refer to the [remote
|
64 |
|
65 |
## 3. To Reproduce this Domain-Adapted MLLM
|
66 |
|
|
|
5 |
base_model:
|
6 |
- Lin-Chen/open-llava-next-llama3-8b
|
7 |
tags:
|
8 |
+
- remote-sensing
|
9 |
---
|
10 |
# Adapting Multimodal Large Language Models to Domains via Post-Training
|
11 |
|
|
|
28 |
|
29 |
instruction = "What's in the image?"
|
30 |
|
31 |
+
model_path='AdaptLLM/remote-sensing-LLaVA-NeXT-Llama3-8B'
|
32 |
|
33 |
# =========================== Do NOT need to modify the following ===============================
|
34 |
# Load the processor
|
|
|
60 |
|
61 |
## 2. To Evaluate Any MLLM on Domain-Specific Benchmarks
|
62 |
|
63 |
+
Refer to the [remote-sensing-VQA-benchmark](https://huggingface.co/datasets/AdaptLLM/remote-sensing-VQA-benchmark) to reproduce our results and evaluate many other MLLMs on domain-specific benchmarks.
|
64 |
|
65 |
## 3. To Reproduce this Domain-Adapted MLLM
|
66 |
|