Improve model card: Add pipeline tag, paper, project, and code links, and abstract
Browse filesThis PR improves the model card for Hermes 4 by:
* Adding the `pipeline_tag: text-generation` metadata, ensuring the model appears correctly in relevant searches on the Hub (https://huggingface.co/models?pipeline_tag=text-generation).
* Adding a prominent link to the Hugging Face Paper page: https://huggingface.co/papers/2508.18255.
* Adding a prominent link to the GitHub repository: https://github.com/NousResearch/Hermes-4.
* Improving the discoverability of the project page by adding a prominent link at the top: https://huggingface.co/collections/NousResearch/hermes-4-collection-68a731bfd452e20816725728.
* Including a dedicated "Abstract (from Paper)" section with the paper's abstract for a quick overview.
README.md
CHANGED
|
@@ -1,7 +1,11 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
| 2 |
language:
|
| 3 |
- en
|
|
|
|
| 4 |
license: apache-2.0
|
|
|
|
| 5 |
tags:
|
| 6 |
- Qwen-3-14B
|
| 7 |
- instruct
|
|
@@ -18,15 +22,11 @@ tags:
|
|
| 18 |
- long context
|
| 19 |
- roleplaying
|
| 20 |
- chat
|
| 21 |
-
base_model:
|
| 22 |
-
- NousResearch/Hermes-4-14B
|
| 23 |
-
library_name: transformers
|
| 24 |
widget:
|
| 25 |
- example_title: Hermes 4
|
| 26 |
messages:
|
| 27 |
- role: system
|
| 28 |
-
content:
|
| 29 |
-
You are Hermes 4, a capable, neutrally-aligned assistant. Prefer concise,
|
| 30 |
correct answers.
|
| 31 |
- role: user
|
| 32 |
content: Explain the difference between BFS and DFS to a new CS student.
|
|
@@ -34,10 +34,18 @@ model-index:
|
|
| 34 |
- name: Hermes-4-Qwen-3-14B
|
| 35 |
results: []
|
| 36 |
---
|
|
|
|
| 37 |
# Hermes 4 — Qwen 3 14B
|
| 38 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 39 |

|
| 40 |
|
|
|
|
|
|
|
|
|
|
| 41 |
## Model Description
|
| 42 |
|
| 43 |
Hermes 4 14B is a frontier, hybrid-mode **reasoning** model based on Qwen 3 14B by Nous Research that is aligned to **you**.
|
|
@@ -51,11 +59,11 @@ Training highlights include a newly synthesized post-training corpus emphasizing
|
|
| 51 |
|
| 52 |
## What’s new vs Hermes 3
|
| 53 |
|
| 54 |
-
-
|
| 55 |
-
-
|
| 56 |
-
-
|
| 57 |
-
-
|
| 58 |
-
-
|
| 59 |
|
| 60 |
## Our Mission: Frontier Capabilities Aligned to You
|
| 61 |
|
|
@@ -127,8 +135,8 @@ Note that you may also simply place tool definitions into the "tools:" field of
|
|
| 127 |
|
| 128 |
The model will then generate tool calls within `<tool_call> {tool_call} </tool_call>` tags, for easy parsing. The tool_call tags are also added tokens, so it makes it easy to parse while streaming! There are also automatic tool parsers built-in to VLLM and SGLang for Hermes, just set the tool parser in VLLM to `hermes` and in SGLang to `qwen25`.
|
| 129 |
## Inference Notes
|
| 130 |
-
-
|
| 131 |
-
-
|
| 132 |
### Transformers example
|
| 133 |
```python
|
| 134 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
@@ -143,6 +151,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
| 143 |
)
|
| 144 |
messages = [
|
| 145 |
{"role":"system","content":"You are Hermes 4. Be concise."},
|
|
|
|
| 146 |
{"role":"user","content":"Summarize CRISPR in 3 sentences."}
|
| 147 |
]
|
| 148 |
inputs = tokenizer.apply_chat_template(
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- NousResearch/Hermes-4-14B
|
| 4 |
language:
|
| 5 |
- en
|
| 6 |
+
library_name: transformers
|
| 7 |
license: apache-2.0
|
| 8 |
+
pipeline_tag: text-generation
|
| 9 |
tags:
|
| 10 |
- Qwen-3-14B
|
| 11 |
- instruct
|
|
|
|
| 22 |
- long context
|
| 23 |
- roleplaying
|
| 24 |
- chat
|
|
|
|
|
|
|
|
|
|
| 25 |
widget:
|
| 26 |
- example_title: Hermes 4
|
| 27 |
messages:
|
| 28 |
- role: system
|
| 29 |
+
content: You are Hermes 4, a capable, neutrally-aligned assistant. Prefer concise,
|
|
|
|
| 30 |
correct answers.
|
| 31 |
- role: user
|
| 32 |
content: Explain the difference between BFS and DFS to a new CS student.
|
|
|
|
| 34 |
- name: Hermes-4-Qwen-3-14B
|
| 35 |
results: []
|
| 36 |
---
|
| 37 |
+
|
| 38 |
# Hermes 4 — Qwen 3 14B
|
| 39 |
|
| 40 |
+
Presented in [Hermes 4 Technical Report](https://huggingface.co/papers/2508.18255).
|
| 41 |
+
**Project Page**: [Hermes 4 Collection](https://huggingface.co/collections/NousResearch/hermes-4-collection-68a731bfd452e20816725728)
|
| 42 |
+
**Code**: [GitHub Repository](https://github.com/NousResearch/Hermes-4)
|
| 43 |
+
|
| 44 |

|
| 45 |
|
| 46 |
+
## Abstract (from Paper)
|
| 47 |
+
We present Hermes 4, a family of hybrid reasoning models that combine structured, multi-turn reasoning with broad instruction-following ability. We describe the challenges encountered during data curation, synthesis, training, and evaluation, and outline the solutions employed to address these challenges at scale. We comprehensively evaluate across mathematical reasoning, coding, knowledge, comprehension, and alignment benchmarks, and we report both quantitative performance and qualitative behavioral analysis. To support open research, all model weights are published publicly at this https URL
|
| 48 |
+
|
| 49 |
## Model Description
|
| 50 |
|
| 51 |
Hermes 4 14B is a frontier, hybrid-mode **reasoning** model based on Qwen 3 14B by Nous Research that is aligned to **you**.
|
|
|
|
| 59 |
|
| 60 |
## What’s new vs Hermes 3
|
| 61 |
|
| 62 |
+
- **Post-training corpus**: Massively increased dataset size from 1M samples and 1.2B tokens to **~5M samples / ~60B tokens** blended across reasoning and non-reasoning data.
|
| 63 |
+
- **Hybrid reasoning mode** with explicit `<think>…</think>` segments when the model decides to deliberate, and options to make your responses faster when you want.
|
| 64 |
+
- **Reasoning** that is top quality, expressive, improves math, code, STEM, logic, and even creative writing and subjective responses.
|
| 65 |
+
- **Schema adherence & structured outputs**: trained to produce valid JSON for given schemas and to repair malformed objects.
|
| 66 |
+
- **Much easier to steer and align**: extreme improvements on steerability, especially on reduced refusal rates.
|
| 67 |
|
| 68 |
## Our Mission: Frontier Capabilities Aligned to You
|
| 69 |
|
|
|
|
| 135 |
|
| 136 |
The model will then generate tool calls within `<tool_call> {tool_call} </tool_call>` tags, for easy parsing. The tool_call tags are also added tokens, so it makes it easy to parse while streaming! There are also automatic tool parsers built-in to VLLM and SGLang for Hermes, just set the tool parser in VLLM to `hermes` and in SGLang to `qwen25`.
|
| 137 |
## Inference Notes
|
| 138 |
+
- **Sampling defaults that work well:** `temperature=0.6, top_p=0.95, top_k=20`.
|
| 139 |
+
- **Template:** Use the ChatML chat format for Hermes 4 14B as shown above, or set `add_generation_prompt=True` when using `tokenizer.apply_chat_template(...)`.
|
| 140 |
### Transformers example
|
| 141 |
```python
|
| 142 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
|
|
| 151 |
)
|
| 152 |
messages = [
|
| 153 |
{"role":"system","content":"You are Hermes 4. Be concise."},
|
| 154 |
+
|
| 155 |
{"role":"user","content":"Summarize CRISPR in 3 sentences."}
|
| 156 |
]
|
| 157 |
inputs = tokenizer.apply_chat_template(
|