@@ -88,6 +88,18 @@ def get_device() -> str:
8888 "qwen-14b-chat-int8" : os .path .join (MODEL_PATH , "Qwen-14B-Chat-Int8" ),
8989 # https://huggingface.co/Qwen/Qwen-14B-Chat-Int4
9090 "qwen-14b-chat-int4" : os .path .join (MODEL_PATH , "Qwen-14B-Chat-Int4" ),
91+ # https://huggingface.co/Qwen/Qwen-72B-Chat
92+ "qwen-72b-chat" : os .path .join (MODEL_PATH , "Qwen-72B-Chat" ),
93+ # https://huggingface.co/Qwen/Qwen-72B-Chat-Int8
94+ "qwen-72b-chat-int8" : os .path .join (MODEL_PATH , "Qwen-72B-Chat-Int8" ),
95+ # https://huggingface.co/Qwen/Qwen-72B-Chat-Int4
96+ "qwen-72b-chat-int4" : os .path .join (MODEL_PATH , "Qwen-72B-Chat-Int4" ),
97+ # https://huggingface.co/Qwen/Qwen-1_8B-Chat
98+ "qwen-1.8b-chat" : os .path .join (MODEL_PATH , "Qwen-1_8B-Chat" ),
99+ # https://huggingface.co/Qwen/Qwen-1_8B-Chat-Int8
100+ "qwen-1.8b-chat-int8" : os .path .join (MODEL_PATH , "wen-1_8B-Chat-Int8" ),
101+ # https://huggingface.co/Qwen/Qwen-1_8B-Chat-Int4
102+ "qwen-1.8b-chat-int4" : os .path .join (MODEL_PATH , "Qwen-1_8B-Chat-Int4" ),
91103 # (Llama2 based) We only support WizardLM-13B-V1.2 for now, which is trained from Llama-2 13b, see https://huggingface.co/WizardLM/WizardLM-13B-V1.2
92104 "wizardlm-13b" : os .path .join (MODEL_PATH , "WizardLM-13B-V1.2" ),
93105 # wget https://huggingface.co/TheBloke/vicuna-13B-v1.5-GGUF/resolve/main/vicuna-13b-v1.5.Q4_K_M.gguf -O models/ggml-model-q4_0.gguf
0 commit comments