From 4baf97bd804c8ea086bc638d2edce4e71c5df940 Mon Sep 17 00:00:00 2001 From: "copilot-swe-agent[bot]" <198982749+Copilot@users.noreply.github.com> Date: Sun, 5 Apr 2026 16:09:22 +0000 Subject: [PATCH] remove dead warnings variable/parameter from audit_report and orphaned comment Agent-Logs-Url: https://github.com/github/gh-aw/sessions/4da4e180-8d54-44b3-8f05-d06682779fd6 Co-authored-by: pelikhan <4175913+pelikhan@users.noreply.github.com> --- pkg/cli/audit_agent_output_test.go | 3 +-- pkg/cli/audit_report.go | 14 ++++---------- pkg/cli/audit_report_analysis.go | 4 ++-- pkg/cli/audit_report_test.go | 31 ++++++++---------------------- 4 files changed, 15 insertions(+), 37 deletions(-) diff --git a/pkg/cli/audit_agent_output_test.go b/pkg/cli/audit_agent_output_test.go index 0439288f4c4..8ac51d330ee 100644 --- a/pkg/cli/audit_agent_output_test.go +++ b/pkg/cli/audit_agent_output_test.go @@ -19,7 +19,6 @@ func TestKeyFindingsGeneration(t *testing.T) { run WorkflowRun metrics MetricsData errors []ErrorInfo - warnings []ErrorInfo mcpFailures []MCPFailureReport missingTools []MissingToolReport expectedCount int @@ -99,7 +98,7 @@ func TestKeyFindingsGeneration(t *testing.T) { MissingTools: tt.missingTools, } - findings := generateFindings(processedRun, tt.metrics, tt.errors, tt.warnings) + findings := generateFindings(processedRun, tt.metrics, tt.errors) if len(findings) < tt.expectedCount { t.Errorf("Expected at least %d findings, got %d", tt.expectedCount, len(findings)) diff --git a/pkg/cli/audit_report.go b/pkg/cli/audit_report.go index 930b30f7059..8e751a3fb5c 100644 --- a/pkg/cli/audit_report.go +++ b/pkg/cli/audit_report.go @@ -311,12 +311,9 @@ func buildAuditData(processedRun ProcessedRun, metrics LogMetrics, mcpToolUsage // Build downloaded files list downloadedFiles := extractDownloadedFiles(run.LogsPath) - // No error/warning extraction since error patterns have been removed - var errors []ErrorInfo - var warnings []ErrorInfo - // For failed workflows where the agent never ran (no agent-stdio.log), // extract errors from step log files to surface the actual failure reason. + var errors []ErrorInfo if run.Conclusion == "failure" && run.LogsPath != "" { if stepErrors := extractPreAgentStepErrors(run.LogsPath); len(stepErrors) > 0 { errors = stepErrors @@ -331,7 +328,7 @@ func buildAuditData(processedRun ProcessedRun, metrics LogMetrics, mcpToolUsage agenticAssessments := buildAgenticAssessments(processedRun, metricsData, toolUsage, createdItems, taskDomain, behaviorFingerprint, overview.AwContext) // Generate key findings - findings := generateFindings(processedRun, metricsData, errors, warnings) + findings := generateFindings(processedRun, metricsData, errors) findings = append(findings, generateAgenticAssessmentFindings(agenticAssessments)...) // Generate recommendations @@ -352,8 +349,8 @@ func buildAuditData(processedRun ProcessedRun, metrics LogMetrics, mcpToolUsage mcpServerHealth := buildMCPServerHealth(mcpToolUsage, processedRun.MCPFailures) if auditReportLog.Enabled() { - auditReportLog.Printf("Built audit data: %d jobs, %d errors, %d warnings, %d tool types, %d findings, %d recommendations", - len(jobs), len(errors), len(warnings), len(toolUsage), len(findings), len(recommendations)) + auditReportLog.Printf("Built audit data: %d jobs, %d errors, %d tool types, %d findings, %d recommendations", + len(jobs), len(errors), len(toolUsage), len(findings), len(recommendations)) } return AuditData{ @@ -383,7 +380,6 @@ func buildAuditData(processedRun ProcessedRun, metrics LogMetrics, mcpToolUsage PolicyAnalysis: processedRun.PolicyAnalysis, RedactedDomainsAnalysis: processedRun.RedactedDomainsAnalysis, Errors: errors, - Warnings: warnings, ToolUsage: toolUsage, MCPToolUsage: mcpToolUsage, CreatedItems: createdItems, @@ -773,5 +769,3 @@ func stripGHALogTimestamps(content string) string { } return strings.Join(stripped, "\n") } - -// renderJSON outputs the audit data as JSON diff --git a/pkg/cli/audit_report_analysis.go b/pkg/cli/audit_report_analysis.go index 832934e4f0f..46e03aeb442 100644 --- a/pkg/cli/audit_report_analysis.go +++ b/pkg/cli/audit_report_analysis.go @@ -11,8 +11,8 @@ import ( ) // generateFindings creates key findings from workflow run data -func generateFindings(processedRun ProcessedRun, metrics MetricsData, errors []ErrorInfo, warnings []ErrorInfo) []Finding { - auditReportLog.Printf("Generating findings: errors=%d, warnings=%d, conclusion=%s", len(errors), len(warnings), processedRun.Run.Conclusion) +func generateFindings(processedRun ProcessedRun, metrics MetricsData, errors []ErrorInfo) []Finding { + auditReportLog.Printf("Generating findings: errors=%d, conclusion=%s", len(errors), processedRun.Run.Conclusion) var findings []Finding run := processedRun.Run diff --git a/pkg/cli/audit_report_test.go b/pkg/cli/audit_report_test.go index 3863070733f..f2d3416e92a 100644 --- a/pkg/cli/audit_report_test.go +++ b/pkg/cli/audit_report_test.go @@ -100,7 +100,6 @@ func TestGenerateFindings(t *testing.T) { processedRun ProcessedRun metrics MetricsData errors []ErrorInfo - warnings []ErrorInfo expectedCount int checkFindings func(t *testing.T, findings []Finding) }{ @@ -119,7 +118,6 @@ func TestGenerateFindings(t *testing.T) { WarningCount: 0, }, errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, expectedCount: 1, // Should have success finding checkFindings: func(t *testing.T, findings []Finding) { assertFindingExists(t, findings, "success", "info", @@ -141,7 +139,6 @@ func TestGenerateFindings(t *testing.T) { WarningCount: 0, }, errors: []ErrorInfo{{Type: "error", Message: "Test error"}}, - warnings: []ErrorInfo{}, expectedCount: 1, // Should have failure finding checkFindings: func(t *testing.T, findings []Finding) { finding := findFindingByCategory(findings, "error") @@ -160,7 +157,6 @@ func TestGenerateFindings(t *testing.T) { }(), metrics: MetricsData{ErrorCount: 1}, errors: []ErrorInfo{{Type: "step_failure", Message: strings.Repeat("x", 500)}}, - warnings: []ErrorInfo{}, expectedCount: 1, checkFindings: func(t *testing.T, findings []Finding) { finding := findFindingByCategory(findings, "error") @@ -180,7 +176,6 @@ func TestGenerateFindings(t *testing.T) { ErrorCount: 1, }, errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, expectedCount: 1, checkFindings: func(t *testing.T, findings []Finding) { finding := findFindingByCategory(findings, "error") @@ -201,7 +196,6 @@ func TestGenerateFindings(t *testing.T) { ErrorCount: 0, // no errors extracted — logs were not available }, errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, expectedCount: 1, checkFindings: func(t *testing.T, findings []Finding) { finding := findFindingByCategory(findings, "error") @@ -226,7 +220,6 @@ func TestGenerateFindings(t *testing.T) { Turns: 20, }, errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, expectedCount: 1, // Timeout finding checkFindings: func(t *testing.T, findings []Finding) { assertFindingContains(t, findings, "performance", "Timeout", @@ -243,7 +236,6 @@ func TestGenerateFindings(t *testing.T) { Turns: 5, }, errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, expectedCount: 1, // High cost finding checkFindings: func(t *testing.T, findings []Finding) { assertFindingExists(t, findings, "cost", "high", @@ -260,7 +252,6 @@ func TestGenerateFindings(t *testing.T) { Turns: 5, }, errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, expectedCount: 1, // Moderate cost finding checkFindings: func(t *testing.T, findings []Finding) { assertFindingExists(t, findings, "cost", "medium", @@ -276,8 +267,7 @@ func TestGenerateFindings(t *testing.T) { TokenUsage: 60000, // > 50000 threshold Turns: 5, }, - errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, + errors: []ErrorInfo{}, checkFindings: func(t *testing.T, findings []Finding) { assertFindingContains(t, findings, "performance", "Token Usage", "High token usage should generate a performance finding") @@ -291,8 +281,7 @@ func TestGenerateFindings(t *testing.T) { metrics: MetricsData{ Turns: 15, // > 10 threshold }, - errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, + errors: []ErrorInfo{}, checkFindings: func(t *testing.T, findings []Finding) { assertFindingContains(t, findings, "performance", "Iterations", "Many iterations should generate a performance finding") @@ -315,7 +304,6 @@ func TestGenerateFindings(t *testing.T) { {Type: "error", Message: "Error 5"}, {Type: "error", Message: "Error 6"}, }, - warnings: []ErrorInfo{}, checkFindings: func(t *testing.T, findings []Finding) { assertFindingContains(t, findings, "error", "Multiple Errors", "Multiple errors should generate an error finding") @@ -333,8 +321,7 @@ func TestGenerateFindings(t *testing.T) { metrics: MetricsData{ Turns: 5, }, - errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, + errors: []ErrorInfo{}, checkFindings: func(t *testing.T, findings []Finding) { assertFindingContains(t, findings, "tooling", "MCP Server", "MCP server failures should generate a tooling finding") @@ -353,8 +340,7 @@ func TestGenerateFindings(t *testing.T) { metrics: MetricsData{ Turns: 5, }, - errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, + errors: []ErrorInfo{}, checkFindings: func(t *testing.T, findings []Finding) { assertFindingContains(t, findings, "tooling", "Tools Not Available", "Missing tools should generate a tooling finding") @@ -374,8 +360,7 @@ func TestGenerateFindings(t *testing.T) { metrics: MetricsData{ Turns: 5, }, - errors: []ErrorInfo{}, - warnings: []ErrorInfo{}, + errors: []ErrorInfo{}, checkFindings: func(t *testing.T, findings []Finding) { assertFindingContains(t, findings, "network", "Blocked", "Firewall blocked requests should generate a network finding") @@ -385,7 +370,7 @@ func TestGenerateFindings(t *testing.T) { for _, tt := range tests { t.Run(tt.name, func(t *testing.T) { - findings := generateFindings(tt.processedRun, tt.metrics, tt.errors, tt.warnings) + findings := generateFindings(tt.processedRun, tt.metrics, tt.errors) if tt.expectedCount > 0 { assert.GreaterOrEqual(t, len(findings), tt.expectedCount, @@ -1074,7 +1059,7 @@ func TestFindingSeverityOrdering(t *testing.T) { {Type: "error", Message: "Error 6"}, } - findings := generateFindings(processedRun, metrics, errors, []ErrorInfo{}) + findings := generateFindings(processedRun, metrics, errors) // Should have critical, high, and medium findings severityCounts := make(map[string]int) @@ -1597,7 +1582,7 @@ func TestGenerateFindingsFirewallWithBlockedDomains(t *testing.T) { fw.SetBlockedDomains([]string{"chatgpt.com"}) pr.FirewallAnalysis = fw - findings := generateFindings(pr, MetricsData{}, nil, nil) + findings := generateFindings(pr, MetricsData{}, nil) var networkFinding *Finding for i := range findings {