Wwayu commited on
Commit
edb1d53
·
verified ·
1 Parent(s): 9ca4d18

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +63 -0
README.md ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: NousResearch/Hermes-4-70B
3
+ language:
4
+ - en
5
+ library_name: transformers
6
+ license: llama3
7
+ pipeline_tag: text-generation
8
+ tags:
9
+ - Llama-3.1
10
+ - instruct
11
+ - finetune
12
+ - reasoning
13
+ - hybrid-mode
14
+ - chatml
15
+ - function calling
16
+ - tool use
17
+ - json mode
18
+ - structured outputs
19
+ - atropos
20
+ - dataforge
21
+ - long context
22
+ - roleplaying
23
+ - chat
24
+ - mlx
25
+ - mlx-my-repo
26
+ widget:
27
+ - example_title: Hermes 4
28
+ messages:
29
+ - role: system
30
+ content: You are Hermes 4, a capable, neutrally-aligned assistant. Prefer concise,
31
+ correct answers.
32
+ - role: user
33
+ content: Explain the difference between BFS and DFS to a new CS student.
34
+ model-index:
35
+ - name: Hermes-4-Llama-3.1-70B
36
+ results: []
37
+ ---
38
+
39
+ # Wwayu/Hermes-4-70B-mlx-4Bit
40
+
41
+ The Model [Wwayu/Hermes-4-70B-mlx-4Bit](https://huggingface.co/Wwayu/Hermes-4-70B-mlx-4Bit) was converted to MLX format from [NousResearch/Hermes-4-70B](https://huggingface.co/NousResearch/Hermes-4-70B) using mlx-lm version **0.26.4**.
42
+
43
+ ## Use with mlx
44
+
45
+ ```bash
46
+ pip install mlx-lm
47
+ ```
48
+
49
+ ```python
50
+ from mlx_lm import load, generate
51
+
52
+ model, tokenizer = load("Wwayu/Hermes-4-70B-mlx-4Bit")
53
+
54
+ prompt="hello"
55
+
56
+ if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None:
57
+ messages = [{"role": "user", "content": prompt}]
58
+ prompt = tokenizer.apply_chat_template(
59
+ messages, tokenize=False, add_generation_prompt=True
60
+ )
61
+
62
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
63
+ ```