|
.*: |
|
wbits: 'None' |
|
model_type: 'None' |
|
groupsize: 'None' |
|
pre_layer: 0 |
|
mode: 'cai-chat' |
|
skip_special_tokens: true |
|
custom_stopping_strings: '' |
|
llama-[0-9]*b-4bit$: |
|
wbits: 4 |
|
model_type: 'llama' |
|
.*-(4bit|int4)-(gr128|128g): |
|
wbits: 4 |
|
groupsize: 128 |
|
.*-(gr128|128g)-(4bit|int4): |
|
wbits: 4 |
|
groupsize: 128 |
|
.*-3bit-(gr128|128g): |
|
wbits: 3 |
|
groupsize: 128 |
|
.*-(gr128|128g)-3bit: |
|
wbits: 3 |
|
groupsize: 128 |
|
.*oasst-sft-1-pythia-12b: |
|
mode: 'instruct' |
|
instruction_template: 'Open Assistant' |
|
.*vicuna: |
|
mode: 'instruct' |
|
instruction_template: 'Vicuna-v0' |
|
.*alpaca: |
|
mode: 'instruct' |
|
instruction_template: 'Alpaca' |
|
.*alpaca-native-4bit: |
|
mode: 'instruct' |
|
instruction_template: 'Alpaca' |
|
wbits: 4 |
|
groupsize: 128 |
|
.*(galactica|oasst): |
|
skip_special_tokens: false |
|
.*dolly-v[0-9]-[0-9]*b: |
|
mode: 'instruct' |
|
instruction_template: 'Alpaca' |
|
skip_special_tokens: false |
|
custom_stopping_strings: '"### End"' |
|
.*koala: |
|
mode: 'instruct' |
|
instruction_template: 'Koala' |
|
.*chatglm: |
|
mode: 'instruct' |
|
instruction_template: 'ChatGLM' |
|
.*llava: |
|
mode: 'instruct' |
|
model_type: 'llama' |
|
instruction_template: 'LLaVA' |
|
custom_stopping_strings: '"\n###"' |
|
.*raven: |
|
mode: 'instruct' |
|
instruction_template: 'RWKV-Raven' |
|
|