Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .github/gallery-agent/agent.go
Original file line number Diff line number Diff line change
Expand Up @@ -146,7 +146,7 @@ func getRealReadme(ctx context.Context, repository string) (string, error) {
return "", err
}

content := newFragment.LastMessage().Content
content := result.LastMessage().Content
return cleanTextContent(content), nil
}

Expand Down
16 changes: 0 additions & 16 deletions core/http/endpoints/localai/mcp.go
Original file line number Diff line number Diff line change
Expand Up @@ -162,11 +162,6 @@ func MCPEndpoint(cl *config.ModelConfigLoader, ml *model.ModelLoader, evaluator
return err
}

f, err = defaultLLM.Ask(ctxWithCancellation, f)
if err != nil {
return err
}

resp := &schema.OpenAIResponse{
ID: id,
Created: created,
Expand Down Expand Up @@ -252,17 +247,6 @@ func MCPEndpoint(cl *config.ModelConfigLoader, ml *model.ModelLoader, evaluator
return
}

// Get final response
f, err = defaultLLM.Ask(ctxWithCancellation, f)
if err != nil {
events <- MCPErrorEvent{
Type: "error",
Message: fmt.Sprintf("Failed to get response: %v", err),
}
ended <- err
return
}

// Stream final assistant response
content := f.LastMessage().Content
events <- MCPAssistantEvent{
Expand Down
38 changes: 0 additions & 38 deletions core/http/endpoints/openresponses/responses.go
Original file line number Diff line number Diff line change
Expand Up @@ -1004,12 +1004,6 @@ func handleBackgroundMCPNonStream(ctx context.Context, store *ResponseStore, res
default:
}

// Get final response
f, err = defaultLLM.Ask(ctx, f)
if err != nil {
return nil, fmt.Errorf("failed to get response: %w", err)
}

// Convert fragment to Open Responses format
fPtr := &f
outputItems := convertCogitoFragmentToORItems(fPtr)
Expand Down Expand Up @@ -1186,21 +1180,6 @@ func handleBackgroundMCPStream(ctx context.Context, store *ResponseStore, respon
default:
}

// Get final response
f, err = defaultLLM.Ask(ctx, f)
if err != nil {
select {
case <-ctx.Done():
ended <- ctx.Err()
case events <- map[string]interface{}{
"type": "error",
"message": fmt.Sprintf("Failed to get response: %v", err),
}:
ended <- err
}
return
}

// Stream final assistant message
content := f.LastMessage().Content
messageID := fmt.Sprintf("msg_%s", uuid.New().String())
Expand Down Expand Up @@ -2580,12 +2559,6 @@ func handleMCPNonStream(c echo.Context, responseID string, createdAt int64, inpu
return sendOpenResponsesError(c, 500, "model_error", fmt.Sprintf("failed to execute tools: %v", err), "")
}

// Get final response
f, err = defaultLLM.Ask(ctx, f)
if err != nil {
return sendOpenResponsesError(c, 500, "model_error", fmt.Sprintf("failed to get response: %v", err), "")
}

// Convert fragment to Open Responses format
fPtr := &f
outputItems := convertCogitoFragmentToORItems(fPtr)
Expand Down Expand Up @@ -2730,17 +2703,6 @@ func handleMCPStream(c echo.Context, responseID string, createdAt int64, input *
return
}

// Get final response
f, err = defaultLLM.Ask(ctx, f)
if err != nil {
events <- map[string]interface{}{
"type": "error",
"message": fmt.Sprintf("Failed to get response: %v", err),
}
ended <- err
return
}

// Stream final assistant message
content := f.LastMessage().Content
messageID := fmt.Sprintf("msg_%s", uuid.New().String())
Expand Down
13 changes: 1 addition & 12 deletions core/services/agent_jobs.go
Original file line number Diff line number Diff line change
Expand Up @@ -27,8 +27,8 @@ import (
"github.com/mudler/LocalAI/pkg/model"
"github.com/mudler/LocalAI/pkg/xsync"
"github.com/mudler/cogito"
"github.com/robfig/cron/v3"
"github.com/mudler/xlog"
"github.com/robfig/cron/v3"
)

// AgentJobService manages agent tasks and job execution
Expand Down Expand Up @@ -894,17 +894,6 @@ func (s *AgentJobService) executeJobInternal(job schema.Job, task schema.Task, c
return fmt.Errorf("failed to execute tools: %w", err)
}

// Get final response
f, err = defaultLLM.Ask(ctx, f)
if err != nil {
job.Status = schema.JobStatusFailed
job.Error = fmt.Sprintf("failed to get response: %v", err)
completedAt := time.Now()
job.CompletedAt = &completedAt
s.jobs.Set(job.ID, job)
return fmt.Errorf("failed to get response: %w", err)
}

// Extract traces from fragment.Status after execution
// This provides complete information about tool calls and results
// We use Status data to supplement/replace callback data for completeness
Expand Down
2 changes: 1 addition & 1 deletion go.mod
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@ require (
github.com/mholt/archiver/v3 v3.5.1
github.com/microcosm-cc/bluemonday v1.0.27
github.com/modelcontextprotocol/go-sdk v1.2.0
github.com/mudler/cogito v0.8.1
github.com/mudler/cogito v0.8.2-0.20260214201734-da0d4ceb2b44
github.com/mudler/edgevpn v0.31.1
github.com/mudler/go-processmanager v0.1.0
github.com/mudler/memory v0.0.0-20251216220809-d1256471a6c2
Expand Down
2 changes: 2 additions & 0 deletions go.sum
Original file line number Diff line number Diff line change
Expand Up @@ -511,6 +511,8 @@ github.com/mr-tron/base58 v1.2.0 h1:T/HDJBh4ZCPbU39/+c3rRvE0uKBQlU27+QI8LJ4t64o=
github.com/mr-tron/base58 v1.2.0/go.mod h1:BinMc/sQntlIE1frQmRFPUoPA1Zkr8VRgBdjWI2mNwc=
github.com/mudler/cogito v0.8.1 h1:66qPJkAMrq/Vo8AC/PvXWuVxYPhi7X2DQuJIilL8+3I=
github.com/mudler/cogito v0.8.1/go.mod h1:6sfja3lcu2nWRzEc0wwqGNu/eCG3EWgij+8s7xyUeQ4=
github.com/mudler/cogito v0.8.2-0.20260214201734-da0d4ceb2b44 h1:joGszpItINnZdoL/0p2077Wz2xnxMGRSRgYN5mS7I4c=
github.com/mudler/cogito v0.8.2-0.20260214201734-da0d4ceb2b44/go.mod h1:6sfja3lcu2nWRzEc0wwqGNu/eCG3EWgij+8s7xyUeQ4=
github.com/mudler/edgevpn v0.31.1 h1:7qegiDWd0kAg6ljhNHxqvp8hbo/6BbzSdbb7/2WZfiY=
github.com/mudler/edgevpn v0.31.1/go.mod h1:ftV5B0nKFzm4R8vR80UYnCb2nf7lxCRgAALxUEEgCf8=
github.com/mudler/go-piper v0.0.0-20241023091659-2494246fd9fc h1:RxwneJl1VgvikiX28EkpdAyL4yQVnJMrbquKospjHyA=
Expand Down
Loading