forked from chigkim/Ollama-MMLU-Pro
-
Notifications
You must be signed in to change notification settings - Fork 0
/
config.toml
31 lines (27 loc) · 1.34 KB
/
config.toml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
# Comment to be included in the beginning of the final report.
comment = ""
[server]
url = "http://localhost:11434/api/generate"
api_key = ""
model = "qwen2.5:32b-instruct-q2_K"
timeout = 600.0
[inference]
# Ssettings below are from evaluate_from_local.py for VLLM on TIGER-AI-Lab/MMLU-Pro
temperature = 0.7
top_p = 0.9
max_tokens = 2048
# The variable {subject} will be replaced with appropriate value in runtime.
system_prompt = "The following are multiple choice questions (with answers) about {subject}. Think step by step and then finish your answer with \"the answer is (X)\" where X is the correct letter choice."
# "multi_chat" inserts COT examples into multi-turn messages. Use for instruct/chat models.
# "no_chat" uses v1/completion api. Use for non-instruct/chat model.
# "single_chat" (from the script for GPT-4O) inserts all the COT examples and question into a single message. Not recommended, use only for legacy compatibility.
style = "multi_chat"
[test]
#示例 categories = ['biology', 'business', 'chemistry', 'computer science', 'economics', 'engineering', 'health', 'history', 'law', 'math', 'philosophy', 'physics', 'psychology', 'other']
categories = ['physics']
parallel = 2
[log]
# Verbosity between 0-2
verbosity = 0
# If true, logs exact prompt sent to the model in the test result files.
log_prompt = true