Jan-v1
Overview
Section titled âOverviewâJan-v1 is a 4B parameter model based on Qwen3-4B-thinking, designed for reasoning and problem-solving tasks. The model achieves 91.1% accuracy on SimpleQA through model scaling and fine-tuning approaches.
Performance
Section titled âPerformanceâSimpleQA Benchmark
Section titled âSimpleQA BenchmarkâJan-v1 demonstrates strong factual question-answering capabilities:

At 91.1% accuracy, Jan-v1 outperforms several larger models on SimpleQA, including Perplexityâs 70B model. This performance represents effective scaling and fine-tuning for a 4B parameter model.
Chat and Creativity Benchmarks
Section titled âChat and Creativity BenchmarksâJan-v1 has been evaluated on conversational and creative tasks:

These benchmarks (EQBench, CreativeWriting, and IFBench) measure the modelâs ability to handle conversational nuance, creative expression, and instruction following.
Requirements
Section titled âRequirementsâ- Memory:
- Minimum: 8GB RAM (with Q4 quantization)
- Recommended: 16GB RAM (with Q8 quantization)
- Hardware: CPU or GPU
- API Support: OpenAI-compatible at localhost:1337
Using Jan-v1
Section titled âUsing Jan-v1âQuick Start
Section titled âQuick Startâ- Download Jan Desktop
- Select Jan-v1 from the model list
- Start chatting - no additional configuration needed

Deployment Options
Section titled âDeployment OptionsâUsing vLLM:
vllm serve janhq/Jan-v1-4B \ --host 0.0.0.0 \ --port 1234 \ --enable-auto-tool-choice \ --tool-call-parser hermesUsing llama.cpp:
llama-server --model jan-v1.gguf \ --host 0.0.0.0 \ --port 1234 \ --jinja \ --no-context-shiftRecommended Parameters
Section titled âRecommended Parametersâtemperature: 0.6top_p: 0.95top_k: 20min_p: 0.0max_tokens: 2048What Jan-v1 Does Well
Section titled âWhat Jan-v1 Does Wellâ- Question Answering: 91.1% accuracy on SimpleQA
- Reasoning Tasks: Built on thinking-optimized base model
- Tool Calling: Supports function calling through hermes parser
- Instruction Following: Reliable response to user instructions
Limitations
Section titled âLimitationsâ- Model Size: 4B parameters limits complex reasoning compared to larger models
- Specialized Tasks: Optimized for Q&A and reasoning, not specialized domains
- Context Window: Standard context limitations apply
Available Formats
Section titled âAvailable FormatsâGGUF Quantizations
Section titled âGGUF Quantizationsâ- Q4_K_M: 2.5 GB - Good balance of size and quality
- Q5_K_M: 2.89 GB - Better quality, slightly larger
- Q6_K: 3.31 GB - Near-full quality
- Q8_0: 4.28 GB - Highest quality quantization
Models Available
Section titled âModels AvailableâTechnical Notes
Section titled âTechnical NotesâCommunity
Section titled âCommunityâ- Discussions: HuggingFace Community
- Support: Available through Jan App at jan.ai