QA-DeBERTa-v3-large-qa_cross_self_attn-binary
This model is a fine-tuned version of microsoft/deberta-v3-large on the saiteki-kai/Beavertails-it dataset. It achieves the following results on the evaluation set:
- Loss: 0.3153
- Accuracy: 0.8620
- Unsafe Precision: 0.8821
- Unsafe Recall: 0.8682
- Unsafe F1: 0.8751
- Unsafe Fpr: 0.1456
- Unsafe Aucpr: 0.9544
- Safe Precision: 0.8378
- Safe Recall: 0.8544
- Safe F1: 0.8460
- Safe Fpr: 0.1318
- Safe Aucpr: 0.9185
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 6e-06
- train_batch_size: 64
- eval_batch_size: 128
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 10
Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Unsafe Precision | Unsafe Recall | Unsafe F1 | Unsafe Fpr | Unsafe Aucpr | Safe Precision | Safe Recall | Safe F1 | Safe Fpr | Safe Aucpr |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 0.3179 | 0.2501 | 2114 | 0.3619 | 0.8397 | 0.9046 | 0.7960 | 0.8468 | 0.1054 | 0.9414 | 0.7775 | 0.8946 | 0.8320 | 0.2040 | 0.8915 |
| 0.345 | 0.5001 | 4228 | 0.3324 | 0.8543 | 0.8684 | 0.8701 | 0.8692 | 0.1654 | 0.9480 | 0.8366 | 0.8346 | 0.8356 | 0.1299 | 0.9061 |
| 0.3071 | 0.7502 | 6342 | 0.3232 | 0.8574 | 0.9018 | 0.8347 | 0.8669 | 0.1141 | 0.9518 | 0.8103 | 0.8859 | 0.8464 | 0.1653 | 0.9134 |
| 0.354 | 1.0002 | 8456 | 0.3182 | 0.8590 | 0.8658 | 0.8834 | 0.8745 | 0.1718 | 0.9528 | 0.8499 | 0.8282 | 0.8389 | 0.1166 | 0.9162 |
| 0.3019 | 1.2503 | 10570 | 0.3201 | 0.8603 | 0.8787 | 0.8688 | 0.8737 | 0.1504 | 0.9528 | 0.8377 | 0.8496 | 0.8436 | 0.1312 | 0.9156 |
| 0.2894 | 1.5004 | 12684 | 0.3243 | 0.8599 | 0.8695 | 0.8805 | 0.8749 | 0.1658 | 0.9535 | 0.8476 | 0.8342 | 0.8409 | 0.1195 | 0.9172 |
| 0.2792 | 1.7504 | 14798 | 0.3153 | 0.8620 | 0.8821 | 0.8682 | 0.8751 | 0.1456 | 0.9544 | 0.8378 | 0.8544 | 0.8460 | 0.1318 | 0.9185 |
| 0.3153 | 2.0005 | 16912 | 0.3153 | 0.8630 | 0.8921 | 0.8575 | 0.8744 | 0.1301 | 0.9553 | 0.8295 | 0.8699 | 0.8492 | 0.1425 | 0.9189 |
| 0.3132 | 2.2505 | 19026 | 0.3150 | 0.8599 | 0.8757 | 0.8721 | 0.8739 | 0.1553 | 0.9542 | 0.8403 | 0.8447 | 0.8425 | 0.1279 | 0.9191 |
| 0.263 | 2.5006 | 21140 | 0.3174 | 0.8590 | 0.8678 | 0.8809 | 0.8743 | 0.1684 | 0.9544 | 0.8477 | 0.8316 | 0.8396 | 0.1191 | 0.9202 |
| 0.2615 | 2.7507 | 23254 | 0.3127 | 0.8610 | 0.8830 | 0.8647 | 0.8738 | 0.1437 | 0.9556 | 0.8346 | 0.8563 | 0.8453 | 0.1353 | 0.9217 |
Framework versions
- Transformers 4.57.3
- Pytorch 2.7.1+cu118
- Datasets 4.4.1
- Tokenizers 0.22.1
- Downloads last month
- 76
Model tree for saiteki-kai/QA-DeBERTa-v3-large-qa_cross_self_attn-binary
Base model
microsoft/deberta-v3-largeEvaluation results
- Accuracy on saiteki-kai/Beavertails-itself-reported0.862