tommymarto commited on
Commit
19ab8e1
1 Parent(s): f47e05d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +39 -20
README.md CHANGED
@@ -1,35 +1,35 @@
1
  ---
2
- license: mit
3
  base_model: dbmdz/bert-base-german-uncased
4
- tags:
5
- - generated_from_trainer
 
6
  model-index:
7
- - name: LernnaviBERT
8
- results: []
9
  ---
10
 
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
 
14
- # LernnaviBERT
15
 
16
- This model is a fine-tuned version of [dbmdz/bert-base-german-uncased](https://huggingface.co/dbmdz/bert-base-german-uncased) on the None dataset.
17
- It achieves the following results on the evaluation set:
18
- - Loss: 0.0060
19
 
20
- ## Model description
 
21
 
22
- More information needed
23
 
24
- ## Intended uses & limitations
25
 
26
- More information needed
27
 
28
- ## Training and evaluation data
29
 
30
- More information needed
31
 
32
- ## Training procedure
 
 
33
 
34
  ### Training hyperparameters
35
 
@@ -37,10 +37,9 @@ The following hyperparameters were used during training:
37
  - learning_rate: 2e-05
38
  - train_batch_size: 16
39
  - eval_batch_size: 16
40
- - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
- - num_epochs: 3.0
44
  - mixed_precision_training: Native AMP
45
 
46
  ### Training results
@@ -52,6 +51,26 @@ The following hyperparameters were used during training:
52
  | 0.0096 | 3.0 | 7215 | 0.0072 |
53
 
54
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
55
  ### Framework versions
56
 
57
  - Transformers 4.37.1
 
1
  ---
2
+ library_name: transformers
3
  base_model: dbmdz/bert-base-german-uncased
4
+ license: mit
5
+ language:
6
+ - de
7
  model-index:
8
+ - name: LernnaviBERT
9
+ results: []
10
  ---
11
 
12
+ # LernnaviBERT Model Card
 
13
 
14
+ LernnaviBERT is finetuning of [German BERT](https://huggingface.co/dbmdz/bert-base-german-uncased) on educational textual data from the Lernnavi Intelligent Tutoring Systems (ITS). It is trained on masked language modeling following the BERT training scheme.
15
 
16
+ ### Model Sources
 
 
17
 
18
+ - **Repository:** [https://github.com/epfl-ml4ed/answer-forecasting](https://github.com/epfl-ml4ed/answer-forecasting)
19
+ - **Paper:** [https://arxiv.org/abs/2405.20079](https://arxiv.org/abs/2405.20079)
20
 
21
+ ### Direct Use
22
 
23
+ Being a fine-tuning of a base BERT model, LernnaviBERT is suitable for all BERT uses, especially in the educational domain in the German language.
24
 
25
+ ### Downstream Use
26
 
27
+ LernnaviBERT has been fine-tuned for [MCQ answering](https://huggingface.co/epfl-ml4ed/MCQBert) and Student Answer Forecasting (like [MCQStudentBertCat](https://huggingface.co/epfl-ml4ed/MCQStudentBertCat) and [MCQStudentBertSum](https://huggingface.co/epfl-ml4ed/MCQStudentBertSum)) as described in [https://arxiv.org/abs/2405.20079](https://arxiv.org/abs/2405.20079)
28
 
 
29
 
30
+ ## Training Details
31
+
32
+ The model was trained on text data from a real-world ITS, Lernnavi, on ~40k text pieces for 3 epochs with a batch size of 16, going from an initial perplexity of 1.21 on Lernnavi data to a final perplexity of 1.01
33
 
34
  ### Training hyperparameters
35
 
 
37
  - learning_rate: 2e-05
38
  - train_batch_size: 16
39
  - eval_batch_size: 16
 
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: linear
42
+ - num_epochs: 3
43
  - mixed_precision_training: Native AMP
44
 
45
  ### Training results
 
51
  | 0.0096 | 3.0 | 7215 | 0.0072 |
52
 
53
 
54
+ ## Citation
55
+
56
+ If you find this useful in your work, please cite our paper
57
+
58
+ ```
59
+ @misc{gado2024student,
60
+ title={Student Answer Forecasting: Transformer-Driven Answer Choice Prediction for Language Learning},
61
+ author={Elena Grazia Gado and Tommaso Martorella and Luca Zunino and Paola Mejia-Domenzain and Vinitra Swamy and Jibril Frej and Tanja Käser},
62
+ year={2024},
63
+ eprint={2405.20079},
64
+ archivePrefix={arXiv},
65
+ }
66
+ ```
67
+
68
+ ```
69
+ Gado, E., Martorella, T., Zunino, L., Mejia-Domenzain, P., Swamy, V., Frej, J., Käser, T. (2024).
70
+ Student Answer Forecasting: Transformer-Driven Answer Choice Prediction for Language Learning.
71
+ In: Proceedings of the Conference on Educational Data Mining (EDM 2024).
72
+ ```
73
+
74
  ### Framework versions
75
 
76
  - Transformers 4.37.1