Update README.md
Browse files
README.md
CHANGED
@@ -77,11 +77,15 @@ European Championship => competitions
|
|
77 |
|
78 |
If you want to use flash attention or increase sequence length, please, check the following code:
|
79 |
```python
|
|
|
|
|
|
|
80 |
model = GLiNER.from_pretrained("knowledgator/gliner-llama-1B-v1.0",
|
81 |
_attn_implementation = 'flash_attention_2',
|
82 |
-
max_len = 2048).to('cuda:0')
|
83 |
```
|
84 |
|
|
|
85 |
### Benchmarks
|
86 |
Below you can see the table with benchmarking results on various named entity recognition datasets:
|
87 |
|
|
|
77 |
|
78 |
If you want to use flash attention or increase sequence length, please, check the following code:
|
79 |
```python
|
80 |
+
from gliner import GLiNER
|
81 |
+
import torch
|
82 |
+
|
83 |
model = GLiNER.from_pretrained("knowledgator/gliner-llama-1B-v1.0",
|
84 |
_attn_implementation = 'flash_attention_2',
|
85 |
+
max_len = 2048).to('cuda:0', dtype=torch.float16)
|
86 |
```
|
87 |
|
88 |
+
|
89 |
### Benchmarks
|
90 |
Below you can see the table with benchmarking results on various named entity recognition datasets:
|
91 |
|