diff --git a/README.md b/README.md index 0cbe24cc..3a7e7eab 100644 --- a/README.md +++ b/README.md @@ -259,7 +259,7 @@ Optionally, you can use the following command-line flags: | Flag | Description | |-------------|-------------| -| `--model_type MODEL_TYPE` | Model type of pre-quantized model. Currently gpt2, gptj, gpt_neox, falcon, llama, mpt, gpt_bigcode, dolly-v2, and replit are supported. | +| `--model_type MODEL_TYPE` | Model type of pre-quantized model. Currently gpt2, gptj, gptneox, falcon, llama, mpt, starcoder (gptbigcode), dollyv2, and replit are supported. | #### AutoGPTQ diff --git a/models/config.yaml b/models/config.yaml index ba12e8bc..624840df 100644 --- a/models/config.yaml +++ b/models/config.yaml @@ -5,7 +5,7 @@ .*(gpt-j|gptj|gpt4all-j|malion-6b|pygway|pygmalion-6b|dolly-v1): model_type: 'gptj' .*(gpt-neox|koalpaca-polyglot|polyglot.*koalpaca|polyglot-ko|polyglot_ko|pythia|stablelm|incite|dolly-v2|polycoder|h2ogpt-oig|h2ogpt-oasst1|h2ogpt-gm): - model_type: 'gpt_neox' + model_type: 'gptneox' .*llama: model_type: 'llama' .*bloom: @@ -17,9 +17,9 @@ .*mpt: model_type: 'mpt' .*(starcoder|starchat): - model_type: 'gpt_bigcode' + model_type: 'starcoder' .*dolly-v2: - model_type: 'dolly-v2' + model_type: 'dollyv2' .*replit: model_type: 'replit' llama-65b-gptq-3bit: diff --git a/modules/loaders.py b/modules/loaders.py index d7bd8d48..08a11ac0 100644 --- a/modules/loaders.py +++ b/modules/loaders.py @@ -290,6 +290,7 @@ loaders_model_types = { "dollyv2" "replit", "starcoder", + "gptbigcode", "falcon" ], }