Text-to-Speech
Safetensors
GGUF
qwen2
audio
speech
speech-language-models
conversational
JohBohU commited on
Commit
12fb9f7
·
verified ·
1 Parent(s): eb3b757

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -36,3 +36,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
  tokenizer_config.json filter=lfs diff=lfs merge=lfs -text
38
  new_tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
  tokenizer_config.json filter=lfs diff=lfs merge=lfs -text
38
  new_tokenizer.json filter=lfs diff=lfs merge=lfs -text
39
+ neutts-BF16.gguf filter=lfs diff=lfs merge=lfs -text
40
+ neutts-Q8-0.gguf filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -3,39 +3,12 @@
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
- "bos_token_id": 41577,
7
- "dtype": "bfloat16",
8
- "eos_token_id": 41579,
9
  "hidden_act": "silu",
10
  "hidden_size": 896,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 4864,
13
- "layer_types": [
14
- "full_attention",
15
- "full_attention",
16
- "full_attention",
17
- "full_attention",
18
- "full_attention",
19
- "full_attention",
20
- "full_attention",
21
- "full_attention",
22
- "full_attention",
23
- "full_attention",
24
- "full_attention",
25
- "full_attention",
26
- "full_attention",
27
- "full_attention",
28
- "full_attention",
29
- "full_attention",
30
- "full_attention",
31
- "full_attention",
32
- "full_attention",
33
- "full_attention",
34
- "full_attention",
35
- "full_attention",
36
- "full_attention",
37
- "full_attention"
38
- ],
39
  "max_position_embeddings": 32768,
40
  "max_window_layers": 21,
41
  "model_type": "qwen2",
@@ -45,10 +18,11 @@
45
  "rms_norm_eps": 1e-06,
46
  "rope_scaling": null,
47
  "rope_theta": 1000000.0,
48
- "sliding_window": null,
49
  "tie_word_embeddings": true,
50
- "transformers_version": "4.56.1",
 
51
  "use_cache": true,
52
  "use_sliding_window": false,
53
- "vocab_size": 107567
54
  }
 
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
 
8
  "hidden_act": "silu",
9
  "hidden_size": 896,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 4864,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  "max_position_embeddings": 32768,
13
  "max_window_layers": 21,
14
  "model_type": "qwen2",
 
18
  "rms_norm_eps": 1e-06,
19
  "rope_scaling": null,
20
  "rope_theta": 1000000.0,
21
+ "sliding_window": 32768,
22
  "tie_word_embeddings": true,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.50.3",
25
  "use_cache": true,
26
  "use_sliding_window": false,
27
+ "vocab_size": 217652
28
  }
generation_config.json CHANGED
@@ -10,5 +10,5 @@
10
  "temperature": 0.7,
11
  "top_k": 20,
12
  "top_p": 0.8,
13
- "transformers_version": "4.56.1"
14
  }
 
10
  "temperature": 0.7,
11
  "top_k": 20,
12
  "top_p": 0.8,
13
+ "transformers_version": "4.50.3"
14
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3fb17d99169d09d52f4933f04ff10431246bb1b365d056cf774cdf4209f829bc
3
- size 1101348824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85c7db53fbe8d62be9bc29a0743661adcb0067552488f185b5f2eb2f1ee4179f
3
+ size 1495893752
neutts-BF16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52f123b3e552408fe8532582bf74c9557f695c148fd118b06546fa87f3d70afc
3
+ size 1503776000
neutts-Q8-0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60f8b30c6f476c2aa12d51de2d169a86aa892766124d08091f0907e3947ed6f
3
+ size 802658528
special_tokens_map.json CHANGED
@@ -1,23 +1,31 @@
1
  {
2
- "bos_token": {
3
- "content": "<|endoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
 
 
 
 
 
 
 
 
9
  "eos_token": {
10
  "content": "<|im_end|>",
11
  "lstrip": false,
12
- "normalized": true,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
  "content": "<|im_end|>",
18
  "lstrip": false,
19
- "normalized": true,
20
  "rstrip": false,
21
  "single_word": false
22
  }
23
- }
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
  "eos_token": {
18
  "content": "<|im_end|>",
19
  "lstrip": false,
20
+ "normalized": false,
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
  "pad_token": {
25
  "content": "<|im_end|>",
26
  "lstrip": false,
27
+ "normalized": false,
28
  "rstrip": false,
29
  "single_word": false
30
  }
31
+ }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:101d08658b465af83cee6e77845a968fd3af759ccb3f0a749741aa02e224e191
3
- size 18184379
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74c466530bd698626a5b6a424d204711c58dfff0a6b3dd8b4dbac1e1e8c9aa87
3
+ size 24140239
tokenizer_config.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5fb19f301cb2698d56b52ebdaef7a22e9b312dfcc4a441a0f9d6d29d34bc5d61
3
- size 12002334
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:364126212a294d794d83036954b0154b925c329411da93e68cdd1addeb4a5bea
3
+ size 12065831