GRM-2.5

logo

Hugging Face License

1. Introduction

GRM-2.5 is a 4B-parameter reasoning model built for general-purpose local AI. It is designed to deliver strong performance across a wide range of tasks while remaining efficient and accessible for local inference.

The model is optimized for structured reasoning, helping it produce more accurate, coherent, and reliable responses on complex problems. GRM-2.5 aims to combine strong reasoning ability, practical usability, and efficient deployment in a compact form factor.

2. Key Capabilities

  • Strong Reasoning for Everyday and Advanced Tasks: GRM-2.5 is built to handle both daily conversations and more demanding reasoning workloads with clarity and consistency.
  • Efficient Local Coding and Agentic Use: Despite its compact size, the model is well suited for code generation, structured problem-solving, and local agent-style workflows.
  • Optimized for Local Deployment: GRM-2.5 is designed for accessible inference across a broad range of hardware, making it a practical choice for users who want capable AI running locally.

3. Performance

GRM-2.5 is designed to be a highly capable option for local AI use across many scenarios. It performs well in complex reasoning tasks, everyday chat, coding, and agentic workflows, while maintaining the efficiency expected from a compact 4B model.

Its focus is not only raw capability, but also practical intelligence: strong reasoning, stable long-context behavior, and usability on consumer hardware.

Downloads last month
37
Safetensors
Model size
5B params
Tensor type
BF16
·
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 2 Ask for provider support

Model tree for OrionLLM/GRM-2.5

Finetuned
Qwen/Qwen3.5-4B
Finetuned
(110)
this model
Quantizations
2 models

Collection including OrionLLM/GRM-2.5