AI & ML interests

Thinking, Agentic, and Research Purpose finetunning and training

Recent Activity

DedeProGames  updated a model about 9 hours ago
OrionLLM/GRM2-3b
DedeProGames  published a model about 9 hours ago
OrionLLM/GRM2-3b
DedeProGames  updated a Space about 17 hours ago
OrionLLM/README
View all activity

DedeProGames 
updated a Space about 17 hours ago
DedeProGames 
posted an update 1 day ago
view post
Post
2328
Can small models program?

Although even if they are reasoning AIs, small AIs cannot create extensive and high-quality code, at least that's what is commonly thought.

We present OrionLLM/NanoCoder-0.6b, an AI with just 600 million parameters based on qwen3-0.6b and trained with the dataset nvidia/OpenCodeReasoning.

While not good at complex code, we observed a significant improvement in code generation (especially in Python code), demonstrating that, when trained correctly, small AIs can, in fact, program.
DedeProGames 
posted an update 3 days ago
view post
Post
2507
Introducing GRM Family, a family of fine-tuned small models from the Qwen2.5 family for Long Cot and General Reasoning and Agentic Tasks.

GRM is available in 7b and 1.5b parameter sizes, these models being significantly relevant for complex tasks or local inference agents.
OrionLLM/GRM-7b
OrionLLM/GRM-1.5b
  • 1 reply
·
DedeProGames 
published a Space 6 days ago