File size: 18,104 Bytes
88d205f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Agent Manager
This module provides the main orchestrator for the Code Review Agent.
It coordinates the review process and manages the state of the application.
"""
import os
import time
import logging
import tempfile
import json
import threading
import concurrent.futures
from datetime import datetime
import gradio as gr
from src.core.language_detector import LanguageDetector
from src.services.code_analyzer import CodeAnalyzer
from src.services.report_generator import ReportGenerator
from src.services.repository_service import RepositoryService
from src.services.security_scanner import SecurityScanner
from src.services.performance_analyzer import PerformanceAnalyzer
logger = logging.getLogger(__name__)
class AgentManager:
"""
Main orchestrator for the Code Review Agent.
This class coordinates the review process, manages the application state,
and provides the interface between the UI and the business logic.
"""
def __init__(self):
"""
Initialize the AgentManager.
"""
# Initialize state management
self.state = {
'repo_url': None,
'progress': {},
'results': {},
'current_step': None
}
# Initialize services
self.language_detector = LanguageDetector()
self.code_analyzer = CodeAnalyzer()
self.report_generator = ReportGenerator()
self.repository_service = RepositoryService()
self.security_scanner = SecurityScanner()
self.performance_analyzer = PerformanceAnalyzer()
self.temp_dir = tempfile.mkdtemp(prefix="code_review_agent_")
logger.info(f"Initialized AgentManager with temp directory: {self.temp_dir}")
def start_review(self, repo_url, github_token=None, selected_languages=None, progress_components=None):
"""
Start the code review process for a GitHub repository.
Args:
repo_url (str): The URL of the GitHub repository to review.
github_token (str, optional): GitHub authentication token for private repositories.
selected_languages (list, optional): List of languages to analyze. If None,
languages will be auto-detected.
progress_components (tuple, optional): Tuple containing (progress_group, overall_progress, status_message, step_progress)
from create_progress_tracker().
Returns:
tuple: (progress_group, overall_progress, status_message, results_dashboard) - Updated UI components.
"""
# Initialize or use provided progress components
if progress_components:
progress_group, overall_progress, status_message, step_progress = progress_components
else:
progress_group = gr.Group(visible=True)
overall_progress = gr.Slider(value=0)
status_message = gr.Markdown("*Starting review...*")
step_progress = {}
try:
# Initialize state
self.state = {
'repo_url': repo_url,
'progress': {},
'results': {},
'current_step': None
}
# Store step progress components
self.step_progress = step_progress
# Clone repository
self._update_progress("Repository Cloning", 0, overall_progress, status_message)
repo_path = self._clone_repository(repo_url, github_token)
self._update_progress("Repository Cloning", 100, overall_progress, status_message)
# Detect languages
self._update_progress("Language Detection", 0, overall_progress, status_message)
if selected_languages and len(selected_languages) > 0:
languages = selected_languages
logger.info(f"Using selected languages: {languages}")
else:
languages = self.language_detector.detect_languages(repo_path)
logger.info(f"Auto-detected languages: {languages}")
self.state['languages'] = languages
self._update_progress("Language Detection", 100, overall_progress, status_message)
# Initialize progress for all steps
self._update_progress("Code Analysis", 0, overall_progress, status_message)
self._update_progress("Security Scanning", 0, overall_progress, status_message)
self._update_progress("Performance Analysis", 0, overall_progress, status_message)
self._update_progress("AI Review", 0, overall_progress, status_message)
# Create a thread lock for updating shared state
lock = threading.Lock()
results = {}
# Define worker functions for each analysis type
def run_code_analysis():
try:
code_results = self.code_analyzer.analyze_repository(repo_path, languages)
with lock:
results['code_analysis'] = code_results
self._update_progress("Code Analysis", 100, overall_progress, status_message)
except Exception as e:
logger.error(f"Error in code analysis thread: {e}")
with lock:
results['code_analysis'] = {'status': 'error', 'error': str(e)}
self._update_progress("Code Analysis", 100, overall_progress, status_message)
def run_security_scan():
try:
security_results = self.security_scanner.scan_repository(repo_path, languages)
with lock:
results['security'] = security_results
self._update_progress("Security Scanning", 100, overall_progress, status_message)
except Exception as e:
logger.error(f"Error in security scanning thread: {e}")
with lock:
results['security'] = {'status': 'error', 'error': str(e)}
self._update_progress("Security Scanning", 100, overall_progress, status_message)
def run_performance_analysis():
try:
perf_results = self.performance_analyzer.analyze_repository(repo_path, languages)
with lock:
results['performance'] = perf_results
self._update_progress("Performance Analysis", 100, overall_progress, status_message)
except Exception as e:
logger.error(f"Error in performance analysis thread: {e}")
with lock:
results['performance'] = {'status': 'error', 'error': str(e)}
self._update_progress("Performance Analysis", 100, overall_progress, status_message)
def run_ai_review():
try:
ai_results = self._perform_ai_review(repo_path, languages)
with lock:
results['ai_review'] = ai_results
self._update_progress("AI Review", 100, overall_progress, status_message)
except Exception as e:
logger.error(f"Error in AI review thread: {e}")
with lock:
results['ai_review'] = {'status': 'error', 'error': str(e)}
self._update_progress("AI Review", 100, overall_progress, status_message)
# Run all analysis tasks in parallel using ThreadPoolExecutor
with concurrent.futures.ThreadPoolExecutor(max_workers=4) as executor:
executor.submit(run_code_analysis)
executor.submit(run_security_scan)
executor.submit(run_performance_analysis)
executor.submit(run_ai_review)
# Wait for all tasks to complete
executor.shutdown(wait=True)
# Update the state with all results
with lock:
self.state['results'].update(results)
# Get repository info
repo_info = self.repository_service.get_repository_info(repo_path)
self.state['results']['repository_info'] = repo_info
# Generate report
self._update_progress("Report Generation", 0, overall_progress, status_message)
repo_name = repo_url.split('/')[-1].replace('.git', '')
report_paths = self.report_generator.generate_report(
repo_name, self.state['results']
)
self.state['report_paths'] = report_paths
self._update_progress("Report Generation", 100, overall_progress, status_message)
# Update results dashboard
results_dashboard = self._create_results_dashboard(self.state['results'])
results_dashboard.visible = True
return progress_group, overall_progress, status_message, results_dashboard
except Exception as e:
logger.exception(f"Error during code review: {e}")
# Update progress components with error
status_message.value = f"*Error: {str(e)}*"
return progress_group, overall_progress, status_message, None
def export_report(self, results_dashboard, export_format):
"""
Export the code review report in the specified format.
Args:
results_dashboard: The results dashboard component.
export_format (str): The format to export the report in ('pdf', 'json', 'html', 'csv').
Returns:
str: The path to the exported file.
"""
try:
if not self.state.get('results'):
logger.warning("No results available to export")
return None
# Get the actual format value from the textbox component
format_value = export_format.value if hasattr(export_format, 'value') else export_format
# Create exports directory if it doesn't exist
exports_dir = os.path.join(os.path.dirname(__file__), '..', '..', 'exports')
os.makedirs(exports_dir, exist_ok=True)
# Generate filename with timestamp
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
repo_name = self.state['repo_url'].split('/')[-1].replace('.git', '')
filename = f"{repo_name}_review_{timestamp}.{format_value}"
filepath = os.path.join(exports_dir, filename)
# Export report in the specified format using report_generator
report_paths = self.report_generator.generate_report(
repo_name, self.state['results'], format_value
)
if format_value in report_paths:
return report_paths[format_value]
else:
logger.warning(f"Unsupported export format: {format_value}")
return None
logger.info(f"Exported report to {filepath}")
return filepath
except Exception as e:
logger.exception(f"Error exporting report: {e}")
return None
def _clone_repository(self, repo_url, github_token=None):
"""
Clone the GitHub repository to a temporary directory.
Args:
repo_url (str): The URL of the GitHub repository to clone.
github_token (str, optional): GitHub authentication token for private repositories.
Returns:
str: The path to the cloned repository.
"""
# Import the repository service here to avoid circular imports
from src.services.repository_service import RepositoryService
# Create a repository service instance
repo_service = RepositoryService(base_temp_dir=self.temp_dir)
# Clone the repository using the service
try:
# If a GitHub token is provided, use it for authentication
if github_token and github_token.strip():
# Modify the URL to include the token for authentication
auth_url = repo_url.replace('https://', f'https://{github_token}@')
repo_path = repo_service.clone_repository(auth_url)
logger.info(f"Cloned repository using GitHub token authentication")
else:
# Clone without authentication (for public repositories)
repo_path = repo_service.clone_repository(repo_url)
logger.info(f"Cloned repository without authentication")
return repo_path
except Exception as e:
logger.error(f"Error cloning repository: {e}")
raise
def _perform_ai_review(self, repo_path, languages):
"""
Perform AI-powered code review with parallel processing.
Args:
repo_path (str): The path to the repository.
languages (list): List of programming languages to analyze.
Returns:
dict: AI review results.
"""
try:
# Import and use the AI review service
from src.mcp.ai_review import AIReviewService
import os
ai_reviewer = AIReviewService()
# Check if AI review is available
if not ai_reviewer.is_available():
logger.warning("AI review service is not available. Please set NEBIUS_API_KEY in environment variables.")
return {
'error': 'AI review service is not available. Please set NEBIUS_API_KEY in environment variables.',
'suggestions': [],
'issues': []
}
# Get all files in the repository
all_files = []
language_extensions = {
'Python': ['.py'],
'JavaScript': ['.js'],
'TypeScript': ['.ts', '.tsx'],
'Java': ['.java'],
'Go': ['.go'],
'Rust': ['.rs']
}
# Create a list of extensions to look for based on selected languages
extensions_to_check = []
for lang in languages:
if lang in language_extensions:
extensions_to_check.extend(language_extensions[lang])
# Find all files with the specified extensions
for root, _, files in os.walk(repo_path):
for file in files:
file_path = os.path.join(root, file)
_, ext = os.path.splitext(file_path)
if ext in extensions_to_check:
all_files.append(file_path)
# Limit the number of files to review to avoid excessive processing
max_files = 20
if len(all_files) > max_files:
logger.warning(f"Too many files to review ({len(all_files)}). Limiting to {max_files} files.")
all_files = all_files[:max_files]
# Process files in parallel
# Pass None for the optional analysis_results parameter
results = ai_reviewer.review_repository(repo_path, all_files, languages, None)
logger.info(f"AI review completed for {len(all_files)} files across {len(languages)} languages")
return results
except Exception as e:
logger.error(f"Error during AI review: {e}")
return {
'error': str(e),
'suggestions': [],
'issues': []
}
def _update_progress(self, step, value, overall_progress=None, status_message=None):
"""Update progress for a specific step and overall progress."""
# Update state
self.state['current_step'] = step
self.state['progress'][step] = value
# Calculate overall progress
total_steps = len(self.state['progress'])
if total_steps > 0:
overall = sum(self.state['progress'].values()) / total_steps
else:
overall = 0
# Update UI components if provided
if overall_progress is not None:
overall_progress.value = overall
if status_message is not None:
status_message.value = f"*Progress update: {step} - {value}% (Overall: {overall:.1f}%)*"
# Update step progress if available
if hasattr(self, 'step_progress') and step in self.step_progress:
self.step_progress[step].value = value
# Log progress
logger.info(f"Progress update: {step} - {value}% (Overall: {overall:.1f}%)")
def _create_results_dashboard(self, report):
"""
Create a results dashboard component for the UI.
Args:
report (dict): The code review report.
Returns:
gr.Tabs: A Gradio results dashboard component.
"""
# Import the create_results_dashboard function from the UI components
from src.ui.components.results_dashboard import create_results_dashboard
# Create a new results dashboard component
results_dashboard = create_results_dashboard()
# Set the visibility to True
results_dashboard.visible = True
# In a full implementation, we would populate the dashboard with data from the report
# For now, we're just returning the empty dashboard component
return results_dashboard |