1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
|
// Copyright 2026 syzkaller project authors. All rights reserved.
// Use of this source code is governed by Apache 2 LICENSE that can be found in the LICENSE file.
package aflow
import (
"errors"
"fmt"
"google.golang.org/genai"
)
// LLMTool acts like a tool for the parent LLM, but is itself implemented as an LLM agent.
// It can have own tools, different from the parent LLM agent.
// It can do complex multi-step research, and provide a concise answer to the parent LLM
// without polluting its context window.
type LLMTool struct {
// Most fields match that of LLMAgent.
// The prompt is not specified here, and is provided by the parent LLM.
Name string
Model string
TaskType TaskType
// Description of the tool exposed to the parent LLM.
Description string
Instruction string
Tools []Tool
agent *LLMAgent
}
type llmToolArgs struct {
Question string `jsonschema:"Question you have."`
}
type llmToolResults struct {
Answer string `jsonschema:"Answer to your question."`
}
func (t *LLMTool) declaration() *genai.FunctionDeclaration {
return &genai.FunctionDeclaration{
Name: t.Name,
Description: t.Description,
ParametersJsonSchema: mustSchemaFor[llmToolArgs](),
ResponseJsonSchema: mustSchemaFor[llmToolResults](),
}
}
func (t *LLMTool) execute(ctx *Context, args map[string]any) (map[string]any, error) {
a, err := convertFromMap[llmToolArgs](args, false, true)
if err != nil {
return nil, err
}
// We temporarily use ctx.state to provide the prompt to the agent,
// and extract the reply.
ctx.state[llmToolPrompt] = a.Question
defer delete(ctx.state, llmToolPrompt)
if err := t.agent.execute(ctx); err != nil {
return nil, err
}
reply, ok := ctx.state[llmToolReply]
if !ok {
return nil, errors.New("state does not contain LLMTool reply")
}
delete(ctx.state, llmToolReply)
return map[string]any{"Answer": reply}, nil
}
const (
llmToolPrompt = "AFLOW_LLMTOOL_PROMPT"
llmToolReply = "AFLOW_LLMTOOL_REPLY"
)
func (t *LLMTool) verify(ctx *verifyContext) {
t.agent = &LLMAgent{
Name: t.Name,
Model: t.Model,
Reply: llmToolReply,
TaskType: t.TaskType,
Instruction: t.Instruction,
Prompt: fmt.Sprintf("{{.%v}}", llmToolPrompt),
Tools: t.Tools,
}
t.agent.verify(ctx)
}
|