| --- |
| license: mit |
| tags: |
| - lora |
| - training |
| - runpod |
| - ai-toolkit |
| --- |
| |
| # AI Trainer - RunPod Serverless |
|
|
| Single-endpoint multi-model LoRA training with all models cached in this repo. |
|
|
| ## RunPod Deployment |
|
|
| **Set Model field to:** `Aloukik21/trainer` |
|
|
| This will cache all models (~240GB) for fast cold starts. |
|
|
| ## Cached Models |
|
|
| | Model Key | Subfolder | Size | |
| |-----------|-----------|------| |
| | flux_dev | flux-dev/ | ~54GB | |
| | flux_schnell | flux-schnell/ | ~54GB | |
| | wan21_14b | wan21-14b/ | ~75GB | |
| | wan22_14b | wan22-14b/ | ~53GB | |
| | qwen_image | qwen-image/ | ~54GB | |
| | accuracy_recovery_adapters | accuracy_recovery_adapters/ | ~3GB | |
| |
| ## API Usage |
| |
| ### List Models |
| ```json |
| {"input": {"action": "list_models"}} |
| ``` |
| |
| ### Train LoRA |
| ```json |
| { |
| "input": { |
| "action": "train", |
| "model": "flux_dev", |
| "params": { |
| "dataset_path": "/workspace/dataset", |
| "output_path": "/workspace/output", |
| "steps": 1000 |
| } |
| } |
| } |
| ``` |
| |
| ### Cleanup (between different models) |
| ```json |
| {"input": {"action": "cleanup"}} |
| ``` |
|
|
| ## Environment Variables |
|
|
| - `HF_TOKEN`: HuggingFace token (required for some gated models) |
|
|
| ## Auto-Cleanup |
|
|
| Handler automatically cleans up GPU memory when switching between different model types. |
|
|