๐Ÿš€ ShweYon-GPT2-Burmese-v1 (แ€€แ€ญแ€ฏแ€›แ€ฝแ€พแ€ฑแ€šแ€ฏแ€”แ€บ)

ShweYon แ€žแ€Šแ€บ แ€™แ€ผแ€”แ€บแ€™แ€ฌแ€˜แ€ฌแ€žแ€ฌแ€…แ€€แ€ฌแ€ธแ€กแ€แ€ฝแ€€แ€บ แ€žแ€ฎแ€ธแ€žแ€”แ€ทแ€บแ€›แ€Šแ€บแ€›แ€ฝแ€šแ€บแ แ€กแ€แ€ผแ€ฑแ€แ€ถแ€™แ€พแ€…แ€แ€„แ€บแ€€แ€ฌ (From Scratch) แ€แ€Šแ€บแ€†แ€ฑแ€ฌแ€€แ€บแ€‘แ€ฌแ€ธแ€žแ€ฑแ€ฌ Decoder-only Transformer แ€™แ€ฑแ€ฌแ€บแ€’แ€šแ€บแ€–แ€ผแ€…แ€บแ€žแ€Šแ€บแ‹ GPT-2 Architecture แ€€แ€ญแ€ฏ แ€กแ€แ€ผแ€ฑแ€แ€ถแ€‘แ€ฌแ€ธแ€•แ€ผแ€ฎแ€ธ แ€™แ€ผแ€”แ€บแ€™แ€ฌแ€…แ€ฌแ€œแ€ฏแ€ถแ€ธแ€•แ€ฑแ€ซแ€„แ€บแ€ธ แ แ€žแ€ญแ€”แ€บแ€ธแ€€แ€ปแ€ฑแ€ฌแ€บแ€•แ€ซแ€แ€„แ€บแ€žแ€ฑแ€ฌ แ€žแ€”แ€ทแ€บแ€…แ€„แ€บแ€•แ€ผแ€ฎแ€ธแ€žแ€ฌแ€ธ Golden Corpus แ€–แ€ผแ€„แ€ทแ€บ แ€œแ€ฑแ€ทแ€€แ€ปแ€„แ€ทแ€บแ€•แ€ฑแ€ธแ€‘แ€ฌแ€ธแ€•แ€ซแ€žแ€Šแ€บแ‹

๐Ÿ“Œ Model Description

  • Developed by: [แ€ฆแ€ธแ€›แ€ฌแ€‡แ€ญแ€”แ€นแ€’แ€ฌแ€”แ€”แ€นแ€’,URajinda]
  • Model type: Causal Language Model
  • Architecture: GPT-2 based (Customized for Myanmar)
  • Language(s): Burmese (แ€™แ€ผแ€”แ€บแ€™แ€ฌแ€˜แ€ฌแ€žแ€ฌ)
  • Tokenizer: Custom Byte-Pair Encoding (BPE) optimized for Myanmar script.

๐ŸŽฏ Intended Use

แ€คแ€™แ€ฑแ€ฌแ€บแ€’แ€šแ€บแ€€แ€ญแ€ฏ แ€กแ€ฑแ€ฌแ€€แ€บแ€•แ€ซแ€œแ€ฏแ€•แ€บแ€„แ€”แ€บแ€ธแ€™แ€ปแ€ฌแ€ธแ€แ€ฝแ€„แ€บ แ€กแ€žแ€ฏแ€ถแ€ธแ€•แ€ผแ€ฏแ€”แ€ญแ€ฏแ€„แ€บแ€žแ€Šแ€บ-

  • แ€™แ€ผแ€”แ€บแ€™แ€ฌแ€…แ€ฌแ€žแ€ฌแ€ธแ€™แ€ปแ€ฌแ€ธ แ€กแ€œแ€ญแ€ฏแ€กแ€œแ€ปแ€ฑแ€ฌแ€€แ€บ แ€›แ€ฑแ€ธแ€žแ€ฌแ€ธแ€แ€ผแ€„แ€บแ€ธ (Text Generation)
  • แ€…แ€ฌแ€€แ€ผแ€ฑแ€ฌแ€„แ€บแ€ธแ€™แ€ปแ€ฌแ€ธ แ€กแ€†แ€ฏแ€ถแ€ธแ€žแ€แ€บแ€•แ€ฑแ€ธแ€แ€ผแ€„แ€บแ€ธ (Sentence Completion)
  • Chatbot แ€”แ€พแ€„แ€ทแ€บ Assistant แ€™แ€ปแ€ฌแ€ธแ€กแ€แ€ฝแ€€แ€บ แ€กแ€แ€ผแ€ฑแ€แ€ถแ€กแ€–แ€ผแ€…แ€บ แ€กแ€žแ€ฏแ€ถแ€ธแ€•แ€ผแ€ฏแ€แ€ผแ€„แ€บแ€ธ (Fine-tuning base)

๐Ÿ“Š Training Details

  • Steps: 8,255 steps
  • Epochs: 5.0
  • Final Loss: [ 3.1]
  • Hardware: NVIDIA T4 GPU (Google Colab)

โš ๏ธ Limitations

แ€คแ€™แ€ฑแ€ฌแ€บแ€’แ€šแ€บแ€žแ€Šแ€บ Base Model แ€–แ€ผแ€…แ€บแ€žแ€ฑแ€ฌแ€€แ€ผแ€ฑแ€ฌแ€„แ€ทแ€บ แ€กแ€แ€ปแ€€แ€บแ€กแ€œแ€€แ€บแ€กแ€™แ€พแ€ฌแ€ธแ€™แ€ปแ€ฌแ€ธ (Hallucinations) แ€•แ€ซแ€แ€„แ€บแ€”แ€ญแ€ฏแ€„แ€บแ€•แ€ซแ€žแ€Šแ€บแ‹ แ€…แ€€แ€ฌแ€ธแ€•แ€ผแ€ฑแ€ฌแ€†แ€ญแ€ฏแ€™แ€พแ€ฏ แ€•แ€ญแ€ฏแ€™แ€ญแ€ฏแ€€แ€ฑแ€ฌแ€„แ€บแ€ธแ€™แ€ฝแ€”แ€บแ€œแ€ฌแ€…แ€ฑแ€›แ€”แ€บ Chat-based Fine-tuning แ€‘แ€•แ€บแ€™แ€ถแ€•แ€ผแ€ฏแ€œแ€ฏแ€•แ€บแ€›แ€”แ€บ แ€œแ€ญแ€ฏแ€กแ€•แ€บแ€•แ€ซแ€žแ€Šแ€บแ‹

๐Ÿ‘จโ€๐Ÿ’ป How to use

from transformers import pipeline

generator = pipeline('text-generation', model='your-username/ShweYon-GPT2-Burmese-v1')
print(generator("แ€™แ€ผแ€”แ€บแ€™แ€ฌแ€”แ€ญแ€ฏแ€„แ€บแ€„แ€ถแ€žแ€Šแ€บ", max_length=50))
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support