-
Notifications
You must be signed in to change notification settings - Fork 0
/
llm_ollama.flow.json
133 lines (133 loc) · 3.82 KB
/
llm_ollama.flow.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
{
"name": "llm_ollama",
"type": "serial",
"description": "Generate a chat completion with a provided model.",
"input": {
"schema": {
"input": {
"type": "object",
"required": [
"messages"
],
"properties": {
"model": {
"type": "string",
"description": "The model name.",
"default": "llama2"
},
"uri": {
"type": "string",
"description": "The model endpoint.",
"default": "http://localhost:11434/api/chat"
},
"messages": {
"type": "array",
"description": "A list of messages comprising the conversation so far.",
"items": {
"type": "object",
"required": [
"role",
"content"
],
"properties": {
"role": {
"type": "string",
"description": "The role of the message author.",
"enum": [
"system",
"user",
"assistant"
],
"default": "user"
},
"content": {
"type": "string",
"description": "The content of the message."
}
}
}
},
"temperature": {
"type": "number",
"description": "What sampling temperature to use, between 0 and 2",
"default": 0.8
},
"stream": {
"type": "boolean",
"description": "If set, partial message deltas will be sent. Tokens will be sent as data-only server-sent events as they become available, with the stream terminated by a data: [DONE] message.",
"default": true
}
}
},
"output": {
"type": "object",
"properties": {
"content": {
"type": "string",
"description": "The message (or chunk if streamed) generated by the model."
}
}
}
},
"tasks": [
{
"name": "chat",
"type": "http",
"input": {
"body": {
"model": "${input.model or \"llama2\"}",
"messages": "${input.messages}",
"options": {
"temperature": "${0.8 if input.temperature == None else input.temperature}"
},
"stream": "${True if input.stream == None else input.stream}"
},
"method": "POST",
"sse_filter": "${jsonencode(jsondecode(data).message)}",
"uri": "${input.uri or \"http://localhost:11434/api/chat\"}"
}
},
{
"name": "status",
"type": "decision",
"input": {
"expression": "${str(chat.status)}",
"cases": {
"200": {
"name": "stream",
"type": "decision",
"input": {
"expression": "${str(isiterator(chat.body))}",
"cases": {
"True": {
"name": "chunks",
"type": "terminate",
"input": {
"output": {
"iterator": "${chat.body}"
}
}
},
"False": {
"name": "response",
"type": "terminate",
"input": {
"output": "${chat.body.message}"
}
}
}
}
}
},
"default": {
"name": "failure",
"type": "terminate",
"input": {
"error": "${chat.body.error}"
}
}
}
}
]
}
}