1
0
mirror of https://github.com/coalaura/whiskr.git synced 2025-09-08 17:06:42 +00:00

2 Commits

Author SHA1 Message Date
Laura
87d33a8d1d much improved search 2025-08-31 00:25:03 +02:00
Laura
c7a2848d05 wip 2025-08-30 15:06:49 +02:00
4 changed files with 152 additions and 62 deletions

View File

@@ -145,11 +145,14 @@ func (r *Request) Parse() (*openrouter.ChatCompletionRequest, error) {
request.Messages = append(request.Messages, openrouter.SystemMessage(prompt)) request.Messages = append(request.Messages, openrouter.SystemMessage(prompt))
} }
if model.Tools && r.Tools.Search && env.Tokens.Exa != "" { if model.Tools && r.Tools.Search && env.Tokens.Exa != "" && r.Iterations > 1 {
request.Tools = GetSearchTools() request.Tools = GetSearchTools()
request.ToolChoice = "auto" request.ToolChoice = "auto"
request.Messages = append(request.Messages, openrouter.SystemMessage(InternalToolsPrompt)) request.Messages = append(
request.Messages,
openrouter.SystemMessage(fmt.Sprintf(InternalToolsPrompt, r.Iterations-1)),
)
} }
for _, message := range r.Messages { for _, message := range r.Messages {
@@ -273,7 +276,7 @@ func HandleChat(w http.ResponseWriter, r *http.Request) {
for iteration := range raw.Iterations { for iteration := range raw.Iterations {
debug("iteration %d of %d", iteration+1, raw.Iterations) debug("iteration %d of %d", iteration+1, raw.Iterations)
if iteration == raw.Iterations-1 { if len(request.Tools) > 0 && iteration == raw.Iterations-1 {
debug("no more tool calls") debug("no more tool calls")
request.Tools = nil request.Tools = nil

122
exa.go
View File

@@ -7,15 +7,17 @@ import (
"fmt" "fmt"
"net/http" "net/http"
"strings" "strings"
"time"
) )
type ExaResult struct { type ExaResult struct {
Title string `json:"title"` Title string `json:"title"`
URL string `json:"url"` URL string `json:"url"`
PublishedDate string `json:"publishedDate"` PublishedDate string `json:"publishedDate,omitempty"`
SiteName string `json:"siteName,omitempty"`
Text string `json:"text"` Summary string `json:"summary,omitempty"`
Summary string `json:"summary"` Highlights []string `json:"highlights,omitempty"`
Text string `json:"text,omitempty"`
} }
type ExaCost struct { type ExaCost struct {
@@ -23,43 +25,20 @@ type ExaCost struct {
} }
type ExaResults struct { type ExaResults struct {
RequestID string `json:"requestId"` RequestID string `json:"requestId"`
Results []ExaResult `json:"results"` SearchType string `json:"resolvedSearchType"`
Cost ExaCost `json:"costDollars"` Results []ExaResult `json:"results"`
} Cost ExaCost `json:"costDollars"`
func (e *ExaResult) String() string {
var (
label string
text string
)
if e.Text != "" {
label = "Text"
text = e.Text
} else if e.Summary != "" {
label = "Summary"
text = e.Summary
}
return fmt.Sprintf(
"Title: %s \nURL: %s \nPublished Date: %s \n%s: %s",
e.Title,
e.URL,
e.PublishedDate,
label,
strings.TrimSpace(text),
)
} }
func (e *ExaResults) String() string { func (e *ExaResults) String() string {
list := make([]string, len(e.Results)) var builder strings.Builder
for i, result := range e.Results { json.NewEncoder(&builder).Encode(map[string]any{
list[i] = result.String() "results": e.Results,
} })
return strings.Join(list, "\n\n---\n\n") return builder.String()
} }
func NewExaRequest(ctx context.Context, path string, data any) (*http.Request, error) { func NewExaRequest(ctx context.Context, path string, data any) (*http.Request, error) {
@@ -100,15 +79,62 @@ func RunExaRequest(req *http.Request) (*ExaResults, error) {
} }
func ExaRunSearch(ctx context.Context, args SearchWebArguments) (*ExaResults, error) { func ExaRunSearch(ctx context.Context, args SearchWebArguments) (*ExaResults, error) {
if args.NumResults <= 0 {
args.NumResults = 6
} else if args.NumResults < 3 {
args.NumResults = 3
} else if args.NumResults >= 12 {
args.NumResults = 12
}
data := map[string]any{ data := map[string]any{
"query": args.Query, "query": args.Query,
"type": "auto", "type": "auto",
"numResults": args.NumResults, "numResults": args.NumResults,
"contents": map[string]any{ }
"summary": map[string]any{
"query": "Summarize this page only with all information directly relevant to answering the user's question: include key facts, numbers, dates, names, definitions, steps, code or commands, and the page's stance or conclusion; omit fluff and unrelated sections.", if len(args.Domains) > 0 {
}, data["includeDomains"] = args.Domains
}
contents := map[string]any{
"summary": map[string]any{},
"highlights": map[string]any{
"numSentences": 2,
"highlightsPerUrl": 3,
}, },
"livecrawl": "preferred",
}
switch args.Intent {
case "news":
data["category"] = "news"
data["numResults"] = max(8, args.NumResults)
data["startPublishedDate"] = daysAgo(30)
case "docs":
contents["subpages"] = 1
contents["subpageTarget"] = []string{"documentation", "changelog", "release notes"}
case "papers":
data["category"] = "research paper"
data["startPublishedDate"] = daysAgo(365 * 2)
case "code":
data["category"] = "github"
contents["subpages"] = 1
contents["subpageTarget"] = []string{"readme", "changelog", "code"}
case "deep_read":
contents["text"] = map[string]any{
"maxCharacters": 8000,
}
}
data["contents"] = contents
switch args.Recency {
case "month":
data["startPublishedDate"] = daysAgo(30)
case "year":
data["startPublishedDate"] = daysAgo(356)
} }
req, err := NewExaRequest(ctx, "/search", data) req, err := NewExaRequest(ctx, "/search", data)
@@ -121,10 +147,16 @@ func ExaRunSearch(ctx context.Context, args SearchWebArguments) (*ExaResults, er
func ExaRunContents(ctx context.Context, args FetchContentsArguments) (*ExaResults, error) { func ExaRunContents(ctx context.Context, args FetchContentsArguments) (*ExaResults, error) {
data := map[string]any{ data := map[string]any{
"urls": args.URLs, "urls": args.URLs,
"summary": map[string]any{},
"highlights": map[string]any{
"numSentences": 2,
"highlightsPerUrl": 3,
},
"text": map[string]any{ "text": map[string]any{
"maxCharacters": 8000, "maxCharacters": 8000,
}, },
"livecrawl": "preferred",
} }
req, err := NewExaRequest(ctx, "/contents", data) req, err := NewExaRequest(ctx, "/contents", data)
@@ -134,3 +166,7 @@ func ExaRunContents(ctx context.Context, args FetchContentsArguments) (*ExaResul
return RunExaRequest(req) return RunExaRequest(req)
} }
func daysAgo(days int) string {
return time.Now().Add(time.Duration(days) * 24 * time.Hour).Format(time.DateOnly)
}

View File

@@ -1 +1,13 @@
You have access to web search tools. Use `search_web` with `query` (string) and `num_results` (1-10) to find current information - when searching for recent/latest information, always include specific dates or years (e.g., "august 2025"). Use `fetch_contents` with `urls` (array) to read full page content from search results or known URLs. Use `github_repository` with `owner` (string) and `repo` (string) to get repository overviews (info, branches, files, README) without cloning. Formulate specific, targeted queries and provide all required parameters. Call only one tool per response. # Tool use
Use at most 1 tool call per turn. You have %d turns with tool calls total.
search_web({query, num_results?, intent?, recency?, domains?})
- Fresh info & citations. Keep query short; add month/year if freshness matters.
- intent: auto|news|docs|papers|code|deep_read (deep_read may include full text).
- num_results: default 6 (3-12); recency: auto|month|year.
fetch_contents({urls})
- Read 1-5 given URLs for exact content/quotes/numbers.
github_repository({owner,repo})
- Quick repo overview + README excerpt.

View File

@@ -11,8 +11,11 @@ import (
) )
type SearchWebArguments struct { type SearchWebArguments struct {
Query string `json:"query"` Query string `json:"query"`
NumResults int `json:"num_results"` NumResults int `json:"num_results,omitempty"`
Intent string `json:"intent,omitempty"`
Recency string `json:"recency,omitempty"`
Domains []string `json:"domains,omitempty"`
} }
type FetchContentsArguments struct { type FetchContentsArguments struct {
@@ -30,40 +33,60 @@ func GetSearchTools() []openrouter.Tool {
Type: openrouter.ToolTypeFunction, Type: openrouter.ToolTypeFunction,
Function: &openrouter.FunctionDefinition{ Function: &openrouter.FunctionDefinition{
Name: "search_web", Name: "search_web",
Description: "Search the web via Exa in auto mode. Returns up to 10 results with short summaries.", Description: "Search the live web (via Exa /search) and return summaries, highlights, and optionally full text for the top results.",
Parameters: map[string]any{ Parameters: map[string]any{
"type": "object", "type": "object",
"required": []string{"query", "num_results"}, "required": []string{"query"},
"properties": map[string]any{ "properties": map[string]any{
"query": map[string]any{ "query": map[string]any{
"type": "string", "type": "string",
"description": "A concise, specific search query in natural language.", "description": "A concise, specific search query in natural language. Include month/year if recency matters (e.g., 'august 2025').",
}, },
"num_results": map[string]any{ "num_results": map[string]any{
"type": "integer", "type": "integer",
"description": "Number of results to return (3-10). Default to 6.", "description": "Number of results to return (3-12). Default is 6.",
"minimum": 3, "minimum": 3,
"maximum": 10, "maximum": 10,
}, },
"intent": map[string]any{
"type": "string",
"enum": []string{"auto", "news", "docs", "papers", "code", "deep_read"},
"description": "Search profile. Use 'news' for breaking topics, 'docs' for official docs/changelogs, 'papers' for research, 'code' for repos, 'deep_read' when you need exact quotes/numbers (adds full text). Default 'auto'.",
},
"recency": map[string]any{
"type": "string",
"enum": []string{"auto", "month", "year", "range"},
"description": "Time filter hint. 'month' ~ last 30 days, 'year' ~ last 365 days. Default 'auto'.",
},
"domains": map[string]any{
"type": "array",
"items": map[string]any{
"type": "string",
},
"description": "Restrict to these domains (e.g., ['europa.eu', 'who.int']).",
},
}, },
"additionalProperties": false, "additionalProperties": false,
}, },
Strict: true,
}, },
}, },
{ {
Type: openrouter.ToolTypeFunction, Type: openrouter.ToolTypeFunction,
Function: &openrouter.FunctionDefinition{ Function: &openrouter.FunctionDefinition{
Name: "fetch_contents", Name: "fetch_contents",
Description: "Fetch page contents for one or more URLs via Exa /contents.", Description: "Fetch and summarize page contents for one or more URLs (via Exa /contents). Use when the user provides specific links.",
Parameters: map[string]any{ Parameters: map[string]any{
"type": "object", "type": "object",
"required": []string{"urls"}, "required": []string{"urls"},
"properties": map[string]any{ "properties": map[string]any{
"urls": map[string]any{ "urls": map[string]any{
"type": "array", "type": "array",
"description": "List of URLs (1..N) to fetch.", "description": "List of URLs to fetch.",
"items": map[string]any{"type": "string"}, "items": map[string]any{
"type": "string",
},
"minItems": 1,
"maxItems": 5,
}, },
}, },
"additionalProperties": false, "additionalProperties": false,
@@ -75,14 +98,14 @@ func GetSearchTools() []openrouter.Tool {
Type: openrouter.ToolTypeFunction, Type: openrouter.ToolTypeFunction,
Function: &openrouter.FunctionDefinition{ Function: &openrouter.FunctionDefinition{
Name: "github_repository", Name: "github_repository",
Description: "Get a quick overview of a GitHub repository without cloning: repo info, up to 20 branches (popular first), top-level files/dirs, and the README.", Description: "Fetch repository metadata and README from GitHub.",
Parameters: map[string]any{ Parameters: map[string]any{
"type": "object", "type": "object",
"required": []string{"owner", "repo"}, "required": []string{"owner", "repo"},
"properties": map[string]any{ "properties": map[string]any{
"owner": map[string]any{ "owner": map[string]any{
"type": "string", "type": "string",
"description": "GitHub username or organization (e.g., 'torvalds').", "description": "Repository owner (e.g., 'torvalds').",
}, },
"repo": map[string]any{ "repo": map[string]any{
"type": "string", "type": "string",
@@ -100,7 +123,7 @@ func GetSearchTools() []openrouter.Tool {
func HandleSearchWebTool(ctx context.Context, tool *ToolCall) error { func HandleSearchWebTool(ctx context.Context, tool *ToolCall) error {
var arguments SearchWebArguments var arguments SearchWebArguments
err := json.Unmarshal([]byte(tool.Args), &arguments) err := ParseAndUpdateArgs(tool, &arguments)
if err != nil { if err != nil {
return err return err
} }
@@ -132,7 +155,7 @@ func HandleSearchWebTool(ctx context.Context, tool *ToolCall) error {
func HandleFetchContentsTool(ctx context.Context, tool *ToolCall) error { func HandleFetchContentsTool(ctx context.Context, tool *ToolCall) error {
var arguments FetchContentsArguments var arguments FetchContentsArguments
err := json.Unmarshal([]byte(tool.Args), &arguments) err := ParseAndUpdateArgs(tool, &arguments)
if err != nil { if err != nil {
return err return err
} }
@@ -164,7 +187,7 @@ func HandleFetchContentsTool(ctx context.Context, tool *ToolCall) error {
func HandleGitHubRepositoryTool(ctx context.Context, tool *ToolCall) error { func HandleGitHubRepositoryTool(ctx context.Context, tool *ToolCall) error {
var arguments GitHubRepositoryArguments var arguments GitHubRepositoryArguments
err := json.Unmarshal([]byte(tool.Args), &arguments) err := ParseAndUpdateArgs(tool, &arguments)
if err != nil { if err != nil {
return err return err
} }
@@ -180,3 +203,19 @@ func HandleGitHubRepositoryTool(ctx context.Context, tool *ToolCall) error {
return nil return nil
} }
func ParseAndUpdateArgs(tool *ToolCall, arguments any) error {
err := json.Unmarshal([]byte(tool.Args), arguments)
if err != nil {
return err
}
b, err := json.Marshal(arguments)
if err != nil {
return err
}
tool.Args = string(b)
return nil
}