| | --- |
| | base_model: allenai/OLMo-1B-hf |
| | library_name: peft |
| | --- |
| | |
| | # OLMo Code Python3 Text-Only Model |
| |
|
| | This is a LoRA adapter fine-tuned on the OLMo-1B model for Python 3 code generation tasks. |
| |
|
| | ## Model Details |
| |
|
| | - **Base Model:** allenai/OLMo-1B-hf |
| | - **Model Type:** LoRA Adapter |
| | - **Task:** Causal Language Modeling for Python 3 code |
| | - **Language:** Python 3 |
| | - **License:** MIT |
| | - **Fine-tuned by:** dipikakhullar |
| |
|
| | ## Model Description |
| |
|
| | This model is a LoRA adapter that has been fine-tuned on Python 3 code data. It extends the capabilities of the base OLMo-1B model specifically for Python code generation tasks. |
| |
|
| | ### LoRA Configuration |
| |
|
| | - **LoRA Type:** LORA |
| | - **LoRA Alpha:** 16 |
| | - **LoRA Dropout:** 0.05 |
| | - **LoRA Rank (r):** 8 |
| | - **Target Modules:** down_proj, q_proj, v_proj, up_proj, k_proj, gate_proj, o_proj |
| | - **Task Type:** CAUSAL_LM |
| |
|
| | ## Uses |
| |
|
| | ### Direct Use |
| |
|
| | This model is intended for Python 3 code generation tasks. It can be used to: |
| | - Generate Python code completions |
| | - Assist with code writing |
| | - Provide code suggestions |
| |
|
| | ### Downstream Use |
| |
|
| | The model can be further fine-tuned for specific Python programming tasks or integrated into code generation applications. |
| |
|
| | ### Out-of-Scope Use |
| |
|
| | This model is specifically designed for Python 3 code generation and may not perform well for: |
| | - Other programming languages |
| | - Natural language tasks |
| | - Non-code related tasks |
| |
|
| | ## How to Get Started with the Model |
| |
|
| | ```python |
| | from peft import PeftModel, PeftConfig |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | |
| | # Load the base model and tokenizer |
| | base_model = AutoModelForCausalLM.from_pretrained("allenai/OLMo-1B-hf") |
| | tokenizer = AutoTokenizer.from_pretrained("allenai/OLMo-1B-hf") |
| | |
| | # Load the LoRA adapter |
| | model = PeftModel.from_pretrained(base_model, "dipikakhullar/olmo-code-python3-text-only") |
| | |
| | # Example usage |
| | prompt = "def fibonacci(n):" |
| | inputs = tokenizer(prompt, return_tensors="pt") |
| | outputs = model.generate(**inputs, max_length=100, temperature=0.7) |
| | print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| | ``` |
| |
|
| | ## Training Details |
| |
|
| | ### Training Data |
| |
|
| | The model was fine-tuned on cleaned Python 3 code data specifically prepared for language model training. |
| |
|
| | ### Training Procedure |
| |
|
| | - **Base Model:** allenai/OLMo-1B-hf |
| | - **Fine-tuning Method:** LoRA (Low-Rank Adaptation) |
| | - **Checkpoint:** checkpoint-6000 |
| |
|
| | ## Model Card Contact |
| |
|
| | - **Author:** dipikakhullar |
| | - **Repository:** https://huggingface.co/dipikakhullar/olmo-code-python3-text-only |
| |
|
| | ## Framework versions |
| |
|
| | - PEFT 0.7.1 |
| | - Transformers |
| |
|