|
1 | 1 | huggingface_mini_db = {
|
2 |
| - "starcoder/15b/base": { |
3 |
| - "backend": "autogptq", |
4 |
| - "model_path": "TheBloke/starcoder-GPTQ", |
5 |
| - "model_class_kwargs": {}, |
6 |
| - "required_memory_mb": 18000, |
7 |
| - "T": 4096, |
8 |
| - "filter_caps": ["completion"], |
9 |
| - "deprecated": True, |
10 |
| - }, |
11 |
| - "starcoder/15b/plus": { |
12 |
| - "backend": "autogptq", |
13 |
| - "model_path": "TheBloke/starcoderplus-GPTQ", |
14 |
| - "model_class_kwargs": {}, |
15 |
| - "required_memory_mb": 18000, |
16 |
| - "T": 4096, |
17 |
| - "filter_caps": ["completion"], |
18 |
| - "deprecated": True, |
19 |
| - }, |
20 |
| - "starchat/15b/beta": { |
21 |
| - "backend": "autogptq", |
22 |
| - "model_path": "TheBloke/starchat-beta-GPTQ", |
23 |
| - "model_class_kwargs": {}, |
24 |
| - "required_memory_mb": 18000, |
25 |
| - "T": 4096, |
26 |
| - "filter_caps": ["chat"], |
27 |
| - "deprecated": True, |
28 |
| - }, |
29 |
| - "starcoder/1b/base": { |
30 |
| - "backend": "transformers", |
31 |
| - "model_path": "smallcloudai/starcoderbase-1b", |
32 |
| - "model_class_kwargs": {}, |
33 |
| - "required_memory_mb": 8000, |
34 |
| - "T": 8192, |
35 |
| - "filter_caps": ["completion", "finetune"], |
36 |
| - "deprecated": True, |
37 |
| - }, |
38 |
| - "starcoder/3b/base": { |
39 |
| - "backend": "transformers", |
40 |
| - "model_path": "smallcloudai/starcoderbase-3b", |
41 |
| - "model_class_kwargs": {}, |
42 |
| - "required_memory_mb": 12000, |
43 |
| - "T": 4096, |
44 |
| - "filter_caps": ["completion", "finetune"], |
45 |
| - "deprecated": True, |
46 |
| - }, |
47 |
| - "starcoder/7b/base": { |
48 |
| - "backend": "transformers", |
49 |
| - "model_path": "smallcloudai/starcoderbase-7b", |
50 |
| - "model_class_kwargs": {}, |
51 |
| - "required_memory_mb": 20000, |
52 |
| - "T": 4096, |
53 |
| - "filter_caps": ["completion", "finetune"], |
54 |
| - "deprecated": True, |
55 |
| - }, |
56 |
| - "wizardcoder/15b": { |
57 |
| - "backend": "autogptq", |
58 |
| - "model_path": "TheBloke/WizardCoder-15B-1.0-GPTQ", |
59 |
| - "model_class_kwargs": {}, |
60 |
| - "required_memory_mb": 18000, |
61 |
| - "T": 4096, |
62 |
| - "filter_caps": ["completion"], |
63 |
| - "deprecated": True, |
64 |
| - }, |
65 |
| - "wizardlm/7b": { |
66 |
| - "backend": "autogptq", |
67 |
| - "model_path": "TheBloke/WizardLM-7B-V1.0-Uncensored-GPTQ", |
68 |
| - "model_class_kwargs": {}, |
69 |
| - "required_memory_mb": 8000, |
70 |
| - "T": 2048, |
71 |
| - "filter_caps": ["chat"], |
72 |
| - "deprecated": True, |
73 |
| - }, |
74 |
| - "wizardlm/13b": { |
75 |
| - "backend": "autogptq", |
76 |
| - "model_path": "TheBloke/WizardLM-13B-V1.1-GPTQ", |
77 |
| - "model_class_kwargs": {}, |
78 |
| - "required_memory_mb": 14000, |
79 |
| - "T": 2048, |
80 |
| - "filter_caps": ["chat"], |
81 |
| - "deprecated": True, |
82 |
| - }, |
83 |
| - "llama2/7b": { |
84 |
| - "backend": "autogptq", |
85 |
| - "model_path": "TheBloke/Llama-2-7b-Chat-GPTQ", |
86 |
| - "model_class_kwargs": {}, |
87 |
| - "required_memory_mb": 8000, |
88 |
| - "T": 2048, |
89 |
| - "filter_caps": ["chat"], |
90 |
| - "deprecated": True, |
91 |
| - }, |
92 |
| - "llama2/13b": { |
93 |
| - "backend": "autogptq", |
94 |
| - "model_path": "TheBloke/Llama-2-13B-chat-GPTQ", |
95 |
| - "model_class_kwargs": {}, |
96 |
| - "required_memory_mb": 14000, |
97 |
| - "T": 2048, |
98 |
| - "filter_caps": ["chat"], |
99 |
| - "deprecated": True, |
100 |
| - }, |
101 |
| - "codellama/7b": { |
102 |
| - "backend": "transformers", |
103 |
| - "model_path": "TheBloke/CodeLlama-7B-fp16", |
104 |
| - "model_class_kwargs": {}, |
105 |
| - "required_memory_mb": 14000, |
106 |
| - "T": 2048, |
107 |
| - "filter_caps": ["completion"], |
108 |
| - "deprecated": True, |
109 |
| - }, |
110 |
| - "wizardlm/30b": { |
111 |
| - "backend": "transformers", |
112 |
| - "model_path": "TheBloke/WizardLM-30B-fp16", |
113 |
| - "model_class_kwargs": { |
114 |
| - "load_in_4bit": True, |
115 |
| - }, |
116 |
| - "T": 2048, |
117 |
| - "filter_caps": ["chat"], |
118 |
| - "deprecated": True, |
119 |
| - }, |
120 | 2 | "deepseek-coder/1.3b/base": {
|
121 | 3 | "backend": "transformers",
|
122 | 4 | "model_path": "deepseek-ai/deepseek-coder-1.3b-base",
|
|
0 commit comments