On Semantic Loss Fine-Tuning Approach for Preventing Model Collapse in Causal Reasoning
Paper • 2605.05438 • Published
Fine-tuned Gemma 270M-IT for d-separation causal reasoning using semantic loss with dynamic lambda scheduling.
from transformers import AutoTokenizer, AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("ludwigw/gemma-dseparation-semantic-v2")
tokenizer = AutoTokenizer.from_pretrained("ludwigw/gemma-dseparation-semantic-v2")
@article{deshmukh2026semantic,
title={On Semantic Loss Fine-Tuning Approach for Preventing Model Collapse in Causal Reasoning},
author={Deshmukh, Pratik and Gupta, Atirek},
journal={arXiv preprint arXiv:2605.05438},
year={2026}
}