german medbert model details (#8266)

* model details

* Apply suggestions from code review

Co-authored-by: Julien Chaumond <chaumond@gmail.com>
This commit is contained in:
smanjil 2020-11-06 09:21:13 +01:00 committed by GitHub
parent 96baaafd34
commit ddeecf08e6
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23

View file

@ -0,0 +1,33 @@
---
language: de
---
# German Medical BERT
This is a fine-tuned model on Medical domain for German language and based on German BERT.
## Overview
**Language model:** bert-base-german-cased
**Language:** German
**Fine-tuning:** Medical articles (diseases, symptoms, therapies, etc..)
**Eval data:** NTS-ICD-10 dataset (Classification)
**Infrastructure:** Gogle Colab
## Details
- We fine-tuned using Pytorch with Huggingface library on Colab GPU.
- With standard parameter settings for fine-tuning as mentioned in original BERT's paper.
- Although had to train for upto 25 epochs for classification.
## Performance (Micro precision, recall and f1 score for multilabel code classification)
![performance](https://raw.githubusercontent.com/smanjil/finetune-lm/master/performance.png)
## Author
Manjil Shrestha: `shresthamanjil21 [at] gmail.com`
Get in touch:
[LinkedIn](https://www.linkedin.com/in/manjil-shrestha-038527b4/)