nielsr HF Staff commited on
Commit
a4eb3fc
·
verified ·
1 Parent(s): 1094801

Improve model card: Add pipeline tag, paper, project, and code links, and abstract

Browse files

This PR improves the model card for Hermes 4 by:

* Adding the `pipeline_tag: text-generation` metadata, ensuring the model appears correctly in relevant searches on the Hub (https://huggingface.co/models?pipeline_tag=text-generation).
* Adding a prominent link to the Hugging Face Paper page: https://huggingface.co/papers/2508.18255.
* Adding a prominent link to the GitHub repository: https://github.com/NousResearch/Hermes-4.
* Improving the discoverability of the project page by adding a prominent link at the top: https://huggingface.co/collections/NousResearch/hermes-4-collection-68a731bfd452e20816725728.
* Including a dedicated "Abstract (from Paper)" section with the paper's abstract for a quick overview.

Files changed (1) hide show
  1. README.md +21 -12
README.md CHANGED
@@ -1,7 +1,11 @@
1
  ---
 
 
2
  language:
3
  - en
 
4
  license: apache-2.0
 
5
  tags:
6
  - Qwen-3-14B
7
  - instruct
@@ -18,15 +22,11 @@ tags:
18
  - long context
19
  - roleplaying
20
  - chat
21
- base_model:
22
- - NousResearch/Hermes-4-14B
23
- library_name: transformers
24
  widget:
25
  - example_title: Hermes 4
26
  messages:
27
  - role: system
28
- content: >-
29
- You are Hermes 4, a capable, neutrally-aligned assistant. Prefer concise,
30
  correct answers.
31
  - role: user
32
  content: Explain the difference between BFS and DFS to a new CS student.
@@ -34,10 +34,18 @@ model-index:
34
  - name: Hermes-4-Qwen-3-14B
35
  results: []
36
  ---
 
37
  # Hermes 4 — Qwen 3 14B
38
 
 
 
 
 
39
  ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/roT9o5bMYBtQziRMlaSDf.jpeg)
40
 
 
 
 
41
  ## Model Description
42
 
43
  Hermes 4 14B is a frontier, hybrid-mode **reasoning** model based on Qwen 3 14B by Nous Research that is aligned to **you**.
@@ -51,11 +59,11 @@ Training highlights include a newly synthesized post-training corpus emphasizing
51
 
52
  ## What’s new vs Hermes 3
53
 
54
- - **Post-training corpus**: Massively increased dataset size from 1M samples and 1.2B tokens to **~5M samples / ~60B tokens** blended across reasoning and non-reasoning data.
55
- - **Hybrid reasoning mode** with explicit `<think>…</think>` segments when the model decides to deliberate, and options to make your responses faster when you want.
56
- - **Reasoning** that is top quality, expressive, improves math, code, STEM, logic, and even creative writing and subjective responses.
57
- - **Schema adherence & structured outputs**: trained to produce valid JSON for given schemas and to repair malformed objects.
58
- - **Much easier to steer and align**: extreme improvements on steerability, especially on reduced refusal rates.
59
 
60
  ## Our Mission: Frontier Capabilities Aligned to You
61
 
@@ -127,8 +135,8 @@ Note that you may also simply place tool definitions into the "tools:" field of
127
 
128
  The model will then generate tool calls within `<tool_call> {tool_call} </tool_call>` tags, for easy parsing. The tool_call tags are also added tokens, so it makes it easy to parse while streaming! There are also automatic tool parsers built-in to VLLM and SGLang for Hermes, just set the tool parser in VLLM to `hermes` and in SGLang to `qwen25`.
129
  ## Inference Notes
130
- - **Sampling defaults that work well:** `temperature=0.6, top_p=0.95, top_k=20`.
131
- - **Template:** Use the ChatML chat format for Hermes 4 14B as shown above, or set `add_generation_prompt=True` when using `tokenizer.apply_chat_template(...)`.
132
  ### Transformers example
133
  ```python
134
  from transformers import AutoTokenizer, AutoModelForCausalLM
@@ -143,6 +151,7 @@ model = AutoModelForCausalLM.from_pretrained(
143
  )
144
  messages = [
145
  {"role":"system","content":"You are Hermes 4. Be concise."},
 
146
  {"role":"user","content":"Summarize CRISPR in 3 sentences."}
147
  ]
148
  inputs = tokenizer.apply_chat_template(
 
1
  ---
2
+ base_model:
3
+ - NousResearch/Hermes-4-14B
4
  language:
5
  - en
6
+ library_name: transformers
7
  license: apache-2.0
8
+ pipeline_tag: text-generation
9
  tags:
10
  - Qwen-3-14B
11
  - instruct
 
22
  - long context
23
  - roleplaying
24
  - chat
 
 
 
25
  widget:
26
  - example_title: Hermes 4
27
  messages:
28
  - role: system
29
+ content: You are Hermes 4, a capable, neutrally-aligned assistant. Prefer concise,
 
30
  correct answers.
31
  - role: user
32
  content: Explain the difference between BFS and DFS to a new CS student.
 
34
  - name: Hermes-4-Qwen-3-14B
35
  results: []
36
  ---
37
+
38
  # Hermes 4 — Qwen 3 14B
39
 
40
+ Presented in [Hermes 4 Technical Report](https://huggingface.co/papers/2508.18255).
41
+ **Project Page**: [Hermes 4 Collection](https://huggingface.co/collections/NousResearch/hermes-4-collection-68a731bfd452e20816725728)
42
+ **Code**: [GitHub Repository](https://github.com/NousResearch/Hermes-4)
43
+
44
  ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/roT9o5bMYBtQziRMlaSDf.jpeg)
45
 
46
+ ## Abstract (from Paper)
47
+ We present Hermes 4, a family of hybrid reasoning models that combine structured, multi-turn reasoning with broad instruction-following ability. We describe the challenges encountered during data curation, synthesis, training, and evaluation, and outline the solutions employed to address these challenges at scale. We comprehensively evaluate across mathematical reasoning, coding, knowledge, comprehension, and alignment benchmarks, and we report both quantitative performance and qualitative behavioral analysis. To support open research, all model weights are published publicly at this https URL
48
+
49
  ## Model Description
50
 
51
  Hermes 4 14B is a frontier, hybrid-mode **reasoning** model based on Qwen 3 14B by Nous Research that is aligned to **you**.
 
59
 
60
  ## What’s new vs Hermes 3
61
 
62
+ - **Post-training corpus**: Massively increased dataset size from 1M samples and 1.2B tokens to **~5M samples / ~60B tokens** blended across reasoning and non-reasoning data.
63
+ - **Hybrid reasoning mode** with explicit `<think>…</think>` segments when the model decides to deliberate, and options to make your responses faster when you want.
64
+ - **Reasoning** that is top quality, expressive, improves math, code, STEM, logic, and even creative writing and subjective responses.
65
+ - **Schema adherence & structured outputs**: trained to produce valid JSON for given schemas and to repair malformed objects.
66
+ - **Much easier to steer and align**: extreme improvements on steerability, especially on reduced refusal rates.
67
 
68
  ## Our Mission: Frontier Capabilities Aligned to You
69
 
 
135
 
136
  The model will then generate tool calls within `<tool_call> {tool_call} </tool_call>` tags, for easy parsing. The tool_call tags are also added tokens, so it makes it easy to parse while streaming! There are also automatic tool parsers built-in to VLLM and SGLang for Hermes, just set the tool parser in VLLM to `hermes` and in SGLang to `qwen25`.
137
  ## Inference Notes
138
+ - **Sampling defaults that work well:** `temperature=0.6, top_p=0.95, top_k=20`.
139
+ - **Template:** Use the ChatML chat format for Hermes 4 14B as shown above, or set `add_generation_prompt=True` when using `tokenizer.apply_chat_template(...)`.
140
  ### Transformers example
141
  ```python
142
  from transformers import AutoTokenizer, AutoModelForCausalLM
 
151
  )
152
  messages = [
153
  {"role":"system","content":"You are Hermes 4. Be concise."},
154
+
155
  {"role":"user","content":"Summarize CRISPR in 3 sentences."}
156
  ]
157
  inputs = tokenizer.apply_chat_template(