Skip to content

Commit 6f795b4

Browse files
committed
1 parent 2bb521b commit 6f795b4

File tree

4 files changed

+105
-1
lines changed

4 files changed

+105
-1
lines changed

api/core/model_runtime/model_providers/siliconflow/llm/_position.yaml

+2
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,8 @@
77
- Qwen/Qwen2.5-Coder-7B-Instruct
88
- Qwen/Qwen2-VL-72B-Instruct
99
- Qwen/Qwen2-1.5B-Instruct
10+
- Qwen/Qwen2.5-72B-Instruct-128K
11+
- Vendor-A/Qwen/Qwen2.5-72B-Instruct
1012
- Pro/Qwen/Qwen2-VL-7B-Instruct
1113
- OpenGVLab/InternVL2-26B
1214
- Pro/OpenGVLab/InternVL2-8B
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,51 @@
1+
model: Qwen/Qwen2.5-72B-Instruct-128K
2+
label:
3+
en_US: Qwen/Qwen2.5-72B-Instruct-128K
4+
model_type: llm
5+
features:
6+
- agent-thought
7+
model_properties:
8+
mode: chat
9+
context_size: 131072
10+
parameter_rules:
11+
- name: temperature
12+
use_template: temperature
13+
- name: max_tokens
14+
use_template: max_tokens
15+
type: int
16+
default: 512
17+
min: 1
18+
max: 4096
19+
help:
20+
zh_Hans: 指定生成结果长度的上限。如果生成结果截断,可以调大该参数。
21+
en_US: Specifies the upper limit on the length of generated results. If the generated results are truncated, you can increase this parameter.
22+
- name: top_p
23+
use_template: top_p
24+
- name: top_k
25+
label:
26+
zh_Hans: 取样数量
27+
en_US: Top k
28+
type: int
29+
help:
30+
zh_Hans: 仅从每个后续标记的前 K 个选项中采样。
31+
en_US: Only sample from the top K options for each subsequent token.
32+
required: false
33+
- name: frequency_penalty
34+
use_template: frequency_penalty
35+
- name: response_format
36+
label:
37+
zh_Hans: 回复格式
38+
en_US: Response Format
39+
type: string
40+
help:
41+
zh_Hans: 指定模型必须输出的格式
42+
en_US: specifying the format that the model must output
43+
required: false
44+
options:
45+
- text
46+
- json_object
47+
pricing:
48+
input: '4.13'
49+
output: '4.13'
50+
unit: '0.000001'
51+
currency: RMB
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,51 @@
1+
model: Vendor-A/Qwen/Qwen2.5-72B-Instruct
2+
label:
3+
en_US: Vendor-A/Qwen/Qwen2.5-72B-Instruct
4+
model_type: llm
5+
features:
6+
- agent-thought
7+
model_properties:
8+
mode: chat
9+
context_size: 32768
10+
parameter_rules:
11+
- name: temperature
12+
use_template: temperature
13+
- name: max_tokens
14+
use_template: max_tokens
15+
type: int
16+
default: 512
17+
min: 1
18+
max: 4096
19+
help:
20+
zh_Hans: 指定生成结果长度的上限。如果生成结果截断,可以调大该参数。
21+
en_US: Specifies the upper limit on the length of generated results. If the generated results are truncated, you can increase this parameter.
22+
- name: top_p
23+
use_template: top_p
24+
- name: top_k
25+
label:
26+
zh_Hans: 取样数量
27+
en_US: Top k
28+
type: int
29+
help:
30+
zh_Hans: 仅从每个后续标记的前 K 个选项中采样。
31+
en_US: Only sample from the top K options for each subsequent token.
32+
required: false
33+
- name: frequency_penalty
34+
use_template: frequency_penalty
35+
- name: response_format
36+
label:
37+
zh_Hans: 回复格式
38+
en_US: Response Format
39+
type: string
40+
help:
41+
zh_Hans: 指定模型必须输出的格式
42+
en_US: specifying the format that the model must output
43+
required: false
44+
options:
45+
- text
46+
- json_object
47+
pricing:
48+
input: '1.00'
49+
output: '1.00'
50+
unit: '0.000001'
51+
currency: RMB

api/core/model_runtime/model_providers/siliconflow/llm/qwen2.5-72b-instruct.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ parameter_rules:
1515
type: int
1616
default: 512
1717
min: 1
18-
max: 8192
18+
max: 4096
1919
help:
2020
zh_Hans: 指定生成结果长度的上限。如果生成结果截断,可以调大该参数。
2121
en_US: Specifies the upper limit on the length of generated results. If the generated results are truncated, you can increase this parameter.

0 commit comments

Comments
 (0)