Commit ·
bc2035f
1
Parent(s): 279de95
Update README.md
Browse files
README.md
CHANGED
|
@@ -35,9 +35,9 @@ By default, RobBERT-2022 has the masked language model head used in training. Th
|
|
| 35 |
|
| 36 |
|
| 37 |
```python
|
| 38 |
-
from transformers import AutoTokenizer,
|
| 39 |
-
tokenizer =
|
| 40 |
-
model =
|
| 41 |
```
|
| 42 |
|
| 43 |
You can then use most of [HuggingFace's BERT-based notebooks](https://huggingface.co/transformers/v4.1.1/notebooks.html) for finetuning RobBERT-2022 on your type of Dutch language dataset.
|
|
|
|
| 35 |
|
| 36 |
|
| 37 |
```python
|
| 38 |
+
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
| 39 |
+
tokenizer = AutoTokenizer.from_pretrained("DTAI-KULeuven/robbert-2022-dutch-base")
|
| 40 |
+
model = AutoModelForSequenceClassification.from_pretrained("DTAI-KULeuven/robbert-2022-dutch-base")
|
| 41 |
```
|
| 42 |
|
| 43 |
You can then use most of [HuggingFace's BERT-based notebooks](https://huggingface.co/transformers/v4.1.1/notebooks.html) for finetuning RobBERT-2022 on your type of Dutch language dataset.
|