tibetan-CS-detector

This model is a fine-tuned version of OMRIDRORI/mbert-tibetan-continual-wylie-final on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 2.8365
  • Accuracy: 0.9388
  • Switch Precision: 0.4980
  • Switch Recall: 0.9130
  • Switch F1: 0.6445
  • True Switches: 138
  • Pred Switches: 253
  • Exact Matches: 122
  • Proximity Matches: 4
  • To Auto Precision: 0.6966
  • To Auto Recall: 0.9254
  • To Allo Precision: 0.3902
  • To Allo Recall: 0.9014
  • True To Auto: 67
  • True To Allo: 71
  • Matched To Auto: 62
  • Matched To Allo: 64

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • gradient_accumulation_steps: 4
  • total_train_batch_size: 16
  • optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 200
  • num_epochs: 35
  • mixed_precision_training: Native AMP
  • label_smoothing_factor: 0.05

Training results

Training Loss Epoch Step Validation Loss Accuracy Switch Precision Switch Recall Switch F1 True Switches Pred Switches Exact Matches Proximity Matches To Auto Precision To Auto Recall To Allo Precision To Allo Recall True To Auto True To Allo Matched To Auto Matched To Allo
6.9424 1.9355 30 3.9697 0.4816 0.0 0.0 0.0 138 8 0 0 0.0 0.0 0.0 0.0 67 71 0 0
4.7989 3.8710 60 3.2594 0.7331 0.0 0.0 0.0 138 1 0 0 0.0 0.0 0.0 0.0 67 71 0 0
9.9599 5.8065 90 3.9145 0.7658 0.5909 0.2826 0.3824 138 66 39 0 0.6786 0.5672 0.1 0.0141 67 71 38 1
7.1635 7.7419 120 4.4059 0.7665 0.3818 0.4565 0.4158 138 165 62 1 0.6438 0.7015 0.1739 0.2254 67 71 47 16
10.5361 9.6774 150 5.7618 0.7737 0.3556 0.6159 0.4509 138 239 82 3 0.6667 0.8358 0.1871 0.4085 67 71 56 29
9.5003 11.6129 180 4.0246 0.8587 0.5741 0.4493 0.5041 138 108 62 0 0.7237 0.8209 0.2188 0.0986 67 71 55 7
11.3652 13.5484 210 3.3524 0.9056 0.4911 0.6014 0.5407 138 169 82 1 0.6818 0.8955 0.2840 0.3239 67 71 60 23
4.7329 15.4839 240 2.6446 0.9111 0.5337 0.6304 0.5781 138 163 85 2 0.6667 0.8955 0.3699 0.3803 67 71 60 27
2.2142 17.4194 270 4.7999 0.9163 0.5 0.8406 0.6270 138 232 114 2 0.6778 0.9104 0.3873 0.7746 67 71 61 55
6.1957 19.3548 300 2.5471 0.9232 0.5928 0.8333 0.6928 138 194 113 2 0.6932 0.9104 0.5094 0.7606 67 71 61 54
6.6179 21.2903 330 2.7181 0.9266 0.5619 0.8551 0.6782 138 210 116 2 0.6977 0.8955 0.4677 0.8169 67 71 60 58
1.6293 23.2258 360 2.1611 0.9365 0.4939 0.8768 0.6319 138 245 118 3 0.6813 0.9254 0.3831 0.8310 67 71 62 59
1.7535 25.1613 390 2.1557 0.9381 0.5105 0.8841 0.6472 138 239 119 3 0.7093 0.9104 0.3987 0.8592 67 71 61 61
1.4616 27.0968 420 3.3062 0.9368 0.4808 0.9058 0.6281 138 260 121 4 0.6966 0.9254 0.3684 0.8873 67 71 62 63
10.5341 29.0323 450 2.8365 0.9388 0.4980 0.9130 0.6445 138 253 122 4 0.6966 0.9254 0.3902 0.9014 67 71 62 64

Framework versions

  • Transformers 4.46.3
  • Pytorch 2.4.1+cu121
  • Datasets 2.0.0
  • Tokenizers 0.20.3
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for levshechter/tibetan-CS-detector

Finetuned
(10)
this model