| { |
| "model_type": "lighttiny", |
| "architectures": [ |
| "LightTinyForCausalLM" |
| ], |
| "name": "lighttiny-4m", |
| "vocab_size": 8000, |
| "hidden_dim": 128, |
| "num_regions": 32, |
| "field_size": 256, |
| "field_depth": 488, |
| "total_configs": 3997696, |
| "total_params": 7102464, |
| "layer3_resonance": { |
| "max_active_ratio": 0.05, |
| "activation_threshold": 0.08, |
| "avg_active_ratio": 0.0, |
| "total_activations": 0 |
| }, |
| "layer4_dynamics": { |
| "coupling_strength": 0.1, |
| "damping": 0.01, |
| "dt": 0.01, |
| "max_evolution_steps": 40, |
| "convergence_threshold": 0.008, |
| "min_coherence": 0.2 |
| }, |
| "bos_token_id": 2, |
| "eos_token_id": 3, |
| "pad_token_id": 0 |
| } |