-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathlm_config.yaml
More file actions
55 lines (44 loc) · 1.58 KB
/
lm_config.yaml
File metadata and controls
55 lines (44 loc) · 1.58 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
# Configuration for language models used in different tasks.
# Each task can specify:
# - model_name: The model to use (required)
# - temperature: Sampling temperature (default: 0.9)
# - predictor_type: DSPy predictor class to use (default: "Predict")
# - api_base: Optional API endpoint override
# - api_key: Optional API key override
# - max_tokens: Optional maximum tokens limit
#
# Note: ensure tasks with input images use multimodal models like llama3.2-vision
default:
model_name: "openrouter/deepseek/deepseek-chat"
temperature: 0.9
predictor_type: "ChainOfThought"
summarization:
model_name: "openrouter/deepseek/deepseek-chat"
predictor_type: "Predict"
document_review:
model_name: "openrouter/deepseek/deepseek-chat"
predictor_type: "ChainOfThought"
section_identification:
model_name: "openrouter/deepseek/deepseek-chat"
predictor_type: "ChainOfThought"
section_review:
model_name: "openrouter/deepseek/deepseek-chat"
predictor_type: "ChainOfThought"
image_caption_extraction:
model_name: "ollama_chat/llama3.2-vision:latest"
predictor_type: "Predict"
caption_analysis:
model_name: "openrouter/deepseek/deepseek-chat"
predictor_type: "ChainOfThought"
caption_combination:
model_name: "ollama_chat/llama3.2-vision:latest"
predictor_type: "Predict"
markdown_segmentation:
model_name: "openrouter/deepseek/deepseek-chat"
predictor_type: "Predict"
storm_writer:
model_name: "openrouter/deepseek/deepseek-chat"
predictor_type: "ChainOfThought"
storm_questions:
model_name: "openrouter/deepseek/deepseek-chat"
predictor_type: "ChainOfThought"