File tree 4 files changed +105
-1
lines changed
api/core/model_runtime/model_providers/siliconflow/llm
4 files changed +105
-1
lines changed Original file line number Diff line number Diff line change 7
7
- Qwen/Qwen2.5-Coder-7B-Instruct
8
8
- Qwen/Qwen2-VL-72B-Instruct
9
9
- Qwen/Qwen2-1.5B-Instruct
10
+ - Qwen/Qwen2.5-72B-Instruct-128K
11
+ - Vendor-A/Qwen/Qwen2.5-72B-Instruct
10
12
- Pro/Qwen/Qwen2-VL-7B-Instruct
11
13
- OpenGVLab/InternVL2-26B
12
14
- Pro/OpenGVLab/InternVL2-8B
Original file line number Diff line number Diff line change
1
+ model : Qwen/Qwen2.5-72B-Instruct-128K
2
+ label :
3
+ en_US : Qwen/Qwen2.5-72B-Instruct-128K
4
+ model_type : llm
5
+ features :
6
+ - agent-thought
7
+ model_properties :
8
+ mode : chat
9
+ context_size : 131072
10
+ parameter_rules :
11
+ - name : temperature
12
+ use_template : temperature
13
+ - name : max_tokens
14
+ use_template : max_tokens
15
+ type : int
16
+ default : 512
17
+ min : 1
18
+ max : 4096
19
+ help :
20
+ zh_Hans : 指定生成结果长度的上限。如果生成结果截断,可以调大该参数。
21
+ en_US : Specifies the upper limit on the length of generated results. If the generated results are truncated, you can increase this parameter.
22
+ - name : top_p
23
+ use_template : top_p
24
+ - name : top_k
25
+ label :
26
+ zh_Hans : 取样数量
27
+ en_US : Top k
28
+ type : int
29
+ help :
30
+ zh_Hans : 仅从每个后续标记的前 K 个选项中采样。
31
+ en_US : Only sample from the top K options for each subsequent token.
32
+ required : false
33
+ - name : frequency_penalty
34
+ use_template : frequency_penalty
35
+ - name : response_format
36
+ label :
37
+ zh_Hans : 回复格式
38
+ en_US : Response Format
39
+ type : string
40
+ help :
41
+ zh_Hans : 指定模型必须输出的格式
42
+ en_US : specifying the format that the model must output
43
+ required : false
44
+ options :
45
+ - text
46
+ - json_object
47
+ pricing :
48
+ input : ' 4.13'
49
+ output : ' 4.13'
50
+ unit : ' 0.000001'
51
+ currency : RMB
Original file line number Diff line number Diff line change
1
+ model : Vendor-A/Qwen/Qwen2.5-72B-Instruct
2
+ label :
3
+ en_US : Vendor-A/Qwen/Qwen2.5-72B-Instruct
4
+ model_type : llm
5
+ features :
6
+ - agent-thought
7
+ model_properties :
8
+ mode : chat
9
+ context_size : 32768
10
+ parameter_rules :
11
+ - name : temperature
12
+ use_template : temperature
13
+ - name : max_tokens
14
+ use_template : max_tokens
15
+ type : int
16
+ default : 512
17
+ min : 1
18
+ max : 4096
19
+ help :
20
+ zh_Hans : 指定生成结果长度的上限。如果生成结果截断,可以调大该参数。
21
+ en_US : Specifies the upper limit on the length of generated results. If the generated results are truncated, you can increase this parameter.
22
+ - name : top_p
23
+ use_template : top_p
24
+ - name : top_k
25
+ label :
26
+ zh_Hans : 取样数量
27
+ en_US : Top k
28
+ type : int
29
+ help :
30
+ zh_Hans : 仅从每个后续标记的前 K 个选项中采样。
31
+ en_US : Only sample from the top K options for each subsequent token.
32
+ required : false
33
+ - name : frequency_penalty
34
+ use_template : frequency_penalty
35
+ - name : response_format
36
+ label :
37
+ zh_Hans : 回复格式
38
+ en_US : Response Format
39
+ type : string
40
+ help :
41
+ zh_Hans : 指定模型必须输出的格式
42
+ en_US : specifying the format that the model must output
43
+ required : false
44
+ options :
45
+ - text
46
+ - json_object
47
+ pricing :
48
+ input : ' 1.00'
49
+ output : ' 1.00'
50
+ unit : ' 0.000001'
51
+ currency : RMB
Original file line number Diff line number Diff line change @@ -15,7 +15,7 @@ parameter_rules:
15
15
type : int
16
16
default : 512
17
17
min : 1
18
- max : 8192
18
+ max : 4096
19
19
help :
20
20
zh_Hans : 指定生成结果长度的上限。如果生成结果截断,可以调大该参数。
21
21
en_US : Specifies the upper limit on the length of generated results. If the generated results are truncated, you can increase this parameter.
You can’t perform that action at this time.
0 commit comments