Model Card for Model ID

Model Details

This is a reproduction of a 3.6 million parameter language model from scratch by following the paper TinyStories: How Small Can Language Models Be and Still Speak Coherent English?. The goal of this project is to demostrate that a very small transformer model, when trained on a simpliefied synthetic dataset, can generate fluent, grammatically correct and consistent short stories.

Model Description

This is the model card of a 🤗 transformers model that has been pushed on the Hub.

  • Developed by: Saurav Prateek
  • Model type: Text Generationg (Transformer - Decoder model)
  • Parameters: 3.65 Million
  • Attention Layers: 8
  • Hidden Dimension: 64
  • Attention Heads per Layer: 16
  • Context Window: 512 tokens
  • Vocab Size: ~50K (GPT-Neo Tokenizer)
  • Learning Rate: 5e-4
  • Language(s) (NLP): English
  • License: MIT

Model Sources [optional]

Training Details

Training Data

The model was trained on the TinyStories dataset, which consist of synthetic short stories generated by GPT-3.5/4. The stories use a restricted vocabulary typical of a 3-year-old child.

Training Procedure

The model was trained from scratch on a NVIDIA T4 GPU for around 3 hours to achieve a loss of 2.17. The model was trained for 0.22 epochs estimating around 55K steps. We used EleutherAI/gpt-neo-125M tokenizer model training and inference.

Training Hyperparameters

  • Training regime:
    • Epochs: 0.22
    • Loss: 2.17
    • GPU: NVIDIA T4
    • Training Steps: 55,000
    • Training Time: ~3 hours

Citation [optional]

Downloads last month
-
Safetensors
Model size
3.65M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train SauravP97/tiny-stories-3M

Paper for SauravP97/tiny-stories-3M