Add files using upload-large-folder tool
Browse files- Qwen3-235B-A22B-UD-IQ1_S.gguf +3 -0
- Qwen3-235B-A22B-UD-IQ2_M.gguf +3 -0
- Qwen3-235B-A22B-UD-Q4_K_XL.gguf +3 -0
- config.json +1 -1
    	
        Qwen3-235B-A22B-UD-IQ1_S.gguf
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:243bf999121284e4bb1866451694c52cb4ca4fc66ac2d21063ce21af63a3b367
         | 
| 3 | 
            +
            size 1153521504
         | 
    	
        Qwen3-235B-A22B-UD-IQ2_M.gguf
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:8878e64324d5e31b5bd650dd6167a0f10f347e14e03906a21d4d0e0d6732b61d
         | 
| 3 | 
            +
            size 1212766048
         | 
    	
        Qwen3-235B-A22B-UD-Q4_K_XL.gguf
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:8053b7be334a98a20d62d77bd446647106ec76b43e6fe3fb5b2ad3b53b8556a0
         | 
| 3 | 
            +
            size 30752254816
         | 
    	
        config.json
    CHANGED
    
    | @@ -11,7 +11,7 @@ | |
| 11 | 
             
              "hidden_act": "silu",
         | 
| 12 | 
             
              "hidden_size": 4096,
         | 
| 13 | 
             
              "initializer_range": 0.02,
         | 
| 14 | 
            -
              "intermediate_size":  | 
| 15 | 
             
              "max_position_embeddings": 40960,
         | 
| 16 | 
             
              "max_window_layers": 94,
         | 
| 17 | 
             
              "mlp_only_layers": [],
         | 
|  | |
| 11 | 
             
              "hidden_act": "silu",
         | 
| 12 | 
             
              "hidden_size": 4096,
         | 
| 13 | 
             
              "initializer_range": 0.02,
         | 
| 14 | 
            +
              "intermediate_size": 12288,
         | 
| 15 | 
             
              "max_position_embeddings": 40960,
         | 
| 16 | 
             
              "max_window_layers": 94,
         | 
| 17 | 
             
              "mlp_only_layers": [],
         | 

