dqnScience-v1

dqnScience-v1 is a 4B-parameter flagship reasoning model designed for deep thinking, scientific problem solving, and complex multi-step reasoning.

Unlike lightweight fast-response models, dqnScience-v1 is built to think longer, reason deeper, and solve harder problemsโ€”often performing far above its size class.

dqnScience Banner


Model Description

  • Model type: Causal Language Model
  • Parameters: 4B
  • Primary use: Scientific reasoning and advanced problem solving
  • Style: Deep, structured, step-by-step reasoning

dqnScience-v1 prioritizes reasoning quality over speed, making it ideal for problems that require careful thought, abstraction, and layered logic.


Intended Uses

Direct Use

  • Solving physics, chemistry, and biology problems
  • Logical and analytical reasoning tasks
  • Multi-step problem solving
  • Conceptual understanding of scientific topics
  • Competitive exam-style questions (college level to moderate)

Key Characteristics

  • Strong multi-step reasoning ability
  • Produces structured and detailed explanations
  • Excels at breaking down complex problems
  • Performs above typical 4B models in reasoning capability
  • Designed for consistency and logical correctness
  • Handles abstract and conceptual questions effectively

Usage

dqnScience-v1 is available in multiple formats:

  • GGUF โ†’ llama.cpp, LM Studio
  • MLX โ†’ optimized for Apple Silicon (coming soon)
  • HF Transformers โ†’ universal compatibility

Training Details

dqnScience-v1 is fine-tuned with a strong focus on reasoning-heavy datasets, emphasizing:

  • Deep chain-of-thought reasoning
  • Scientific and logical problem solving
  • Conceptual clarity over memorization
  • Robust multi-step inference

Limitations

  • Slower than lightweight models due to deeper reasoning
  • May over-explain simple questions
  • Not optimized for casual or short-form responses
  • Performance may vary on highly specialized or research-level topics

Efficiency

Despite its strong reasoning capabilities, dqnScience-v1 is optimized to run moderately efficiently on consumer hardware, with support for quantized formats.


License

Apache 2.0


Author

Developed by DQN Labs.

Special thanks to Ram2 for quantization.

This model card was generated with the help of dqnGPT v1.

Downloads last month
826
GGUF
Model size
4B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support