#!/usr/bin/env python3 """ BuddAI Executive v3.1 - Modular Builder BuddAI Executive v3.2 - Hardened Modular Builder Breaks complex tasks into manageable chunks Author: James Gilbert License: MIT """ import sys import os import json import logging import sqlite3 from datetime import datetime from pathlib import Path import http.client import re # noqa: F401 from typing import Optional, List, Dict, Tuple, Union, Generator import zipfile import shutil import queue import socket import argparse import io from urllib.parse import urlparse try: import qrcode except ImportError: qrcode = None try: import psutil except ImportError: psutil = None # Server dependencies try: from fastapi import FastAPI, UploadFile, File, Header, WebSocket, WebSocketDisconnect, Request from fastapi.middleware.cors import CORSMiddleware from fastapi.staticfiles import StaticFiles from fastapi.responses import FileResponse, HTMLResponse, JSONResponse, Response from pydantic import BaseModel import uvicorn SERVER_AVAILABLE = True except ImportError: SERVER_AVAILABLE = False # Configuration OLLAMA_HOST = os.getenv("OLLAMA_HOST", "127.0.0.1") OLLAMA_PORT = int(os.getenv("OLLAMA_PORT", "11434")) DATA_DIR = Path(__file__).parent / "data" DB_PATH = DATA_DIR / "conversations.db" # Validation Config MAX_FILE_SIZE = 50 * 1024 * 1024 # 50MB MAX_UPLOAD_FILES = 10 ALLOWED_TYPES = [ "application/zip", "application/x-zip-compressed", "application/octet-stream", "text/plain", "text/x-python", "text/javascript", "application/javascript", "text/html", "text/css", "text/x-c", "text/x-c++src" ] # Models MODELS = { "fast": "qwen2.5-coder:1.5b", "balanced": "qwen2.5-coder:3b" } # Complexity triggers - if matched, break down the task COMPLEX_TRIGGERS = [ "complete", "entire", "full", "build entire", "build complete", "with ble and", "with servo and", "including", "all of" ] # Module patterns we can detect MODULE_PATTERNS = { "ble": ["bluetooth", "ble", "wireless"], "servo": ["servo", "flipper", "weapon"], "motor": ["motor", "drive", "movement", "l298n"], "safety": ["safety", "timeout", "failsafe", "emergency"], "battery": ["battery", "voltage", "power monitor"], "sensor": ["sensor", "distance", "proximity"] } # --- Connection Pooling --- class OllamaConnectionPool: def __init__(self, host: str, port: int, max_size: int = 10): self.host = host self.port = port self.pool: queue.Queue = queue.Queue(maxsize=max_size) def get_connection(self) -> http.client.HTTPConnection: try: return self.pool.get_nowait() except queue.Empty: return http.client.HTTPConnection(self.host, self.port, timeout=90) def return_connection(self, conn: http.client.HTTPConnection): try: self.pool.put_nowait(conn) except queue.Full: conn.close() OLLAMA_POOL = OllamaConnectionPool(OLLAMA_HOST, OLLAMA_PORT) # --- Shadow Suggestion Engine --- class ShadowSuggestionEngine: """Proactively suggests modules/settings based on user/project history.""" def __init__(self, db_path: Path, user_id: str = "default"): self.db_path = db_path self.user_id = user_id def lookup_recent_module_usage(self, module: str, limit: int = 5) -> List[Tuple[str, str, str]]: """Look up recent usage patterns for a module from repo_index.""" conn = sqlite3.connect(self.db_path) cursor = conn.cursor() cursor.execute( """ SELECT file_path, content, last_modified FROM repo_index WHERE (function_name LIKE ? OR file_path LIKE ?) AND user_id = ? ORDER BY last_modified DESC LIMIT ? """, (f"%{module}%", f"%{module}%", self.user_id, limit) ) results = cursor.fetchall() conn.close() return results def suggest_for_module(self, module: str) -> Optional[str]: """Return a proactive suggestion string for a module if pattern detected.""" history = self.lookup_recent_module_usage(module) if not history: return None # Example: For 'motor', look for L298N and PWM frequency l298n_count = 0 pwm_freqs = [] for _, content, _ in history: if "L298N" in content or "l298n" in content: l298n_count += 1 pwm_matches = re.findall(r'PWM_FREQ\s*=\s*(\d+)', content) pwm_freqs.extend([int(f) for f in pwm_matches]) # Also look for explicit frequency in analogWrite or ledcSetup freq_matches = re.findall(r'(?:ledcSetup|analogWrite)\s*\([^,]+,\s*[^,]+,\s*(\d+)\)', content) pwm_freqs.extend([int(f) for f in freq_matches if f.isdigit()]) if l298n_count >= 2: freq = max(set(pwm_freqs), key=pwm_freqs.count) if pwm_freqs else 500 return f"I see you usually use the L298N with a {freq}Hz PWM frequency on the ESP32-C3. Should I prep that module?" return None def get_proactive_suggestion(self, user_input: str) -> Optional[str]: """ V3.0 Proactive Hook: 1. Identify "Concept" (e.g., 'flipper') 2. Query repo_index for James's most frequent companion modules 3. If 'flipper' often appears with 'safety_timeout', suggest it. """ # 1. Identify Concepts input_lower = user_input.lower() detected_modules = [] for module, keywords in MODULE_PATTERNS.items(): if any(kw in input_lower for kw in keywords): detected_modules.append(module) if not detected_modules: return None # 2. Query repo_index for correlations conn = sqlite3.connect(self.db_path) cursor = conn.cursor() suggestions = [] for module in detected_modules: # Find files containing this module (simple heuristic) cursor.execute("SELECT content FROM repo_index WHERE content LIKE ? AND user_id = ? LIMIT 10", (f"%{module}%", self.user_id)) rows = cursor.fetchall() if not rows: continue # Check for companion modules companions = {} for (content,) in rows: content_lower = content.lower() for other_mod, other_kws in MODULE_PATTERNS.items(): if other_mod != module and other_mod not in detected_modules: if any(kw in content_lower for kw in other_kws): companions[other_mod] = companions.get(other_mod, 0) + 1 # 3. Suggest if frequent (>50% correlation in sample) for other_mod, count in companions.items(): if count >= len(rows) * 0.5: suggestions.append(f"I noticed '{module}' often appears with '{other_mod}' in your repos. Want to include that?") conn.close() return " ".join(list(set(suggestions))) if suggestions else None def get_all_suggestions(self, user_input: str, generated_code: str) -> List[str]: """Aggregate all proactive suggestions into a list.""" suggestions = [] # 1. Companion Modules companion = self.get_proactive_suggestion(user_input) if companion: suggestions.append(companion) # 2. Module Settings input_lower = user_input.lower() for module, keywords in MODULE_PATTERNS.items(): if any(kw in input_lower for kw in keywords): s = self.suggest_for_module(module) if s: suggestions.append(s) # 3. Forge Theory Check if ("motor" in input_lower or "servo" in input_lower) and "applyForge" not in generated_code: suggestions.append("Apply Forge Theory smoothing to movement?") # 4. Safety Check (L298N) if "L298N" in generated_code and "safety" not in generated_code.lower(): suggestions.append("Drive system lacks safety timeout (GilBot_V2 uses 5s failsafe). Add that?") return suggestions class BuddAI: """Executive with task breakdown""" def is_search_query(self, message: str) -> bool: """Check if this is a search query that should query repo_index""" message_lower = message.lower() search_triggers = [ "show me", "find", "search for", "list all", "what functions", "which repos", "do i have", "where did i", "have i used", "examples of", "show all", "display" ] return any(trigger in message_lower for trigger in search_triggers) def search_repositories(self, query: str) -> str: """Search repo_index for relevant functions and code""" conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute("SELECT COUNT(*) FROM repo_index WHERE user_id = ?", (self.user_id,)) count = cursor.fetchone()[0] print(f"\nšŸ” Searching {count} indexed functions...\n") # Extract keywords from query keywords = re.findall(r'\b\w{4,}\b', query.lower()) # Add specific search terms specific_terms = [] if "exponential" in query.lower() or "decay" in query.lower(): specific_terms.append("applyForge") specific_terms.append("exp(") if "forge" in query.lower(): specific_terms.append("Forge") keywords.extend(specific_terms) if not keywords: print("āŒ No search terms found") conn.close() return "No search terms provided." # Build parameterized query conditions = [] params = [] for keyword in keywords: conditions.append("(function_name LIKE ? OR content LIKE ? OR repo_name LIKE ?)") params.extend([f"%{keyword}%", f"%{keyword}%", f"%{keyword}%"]) sql = f"SELECT repo_name, file_path, function_name, content FROM repo_index WHERE ({' OR '.join(conditions)}) AND user_id = ? ORDER BY last_modified DESC LIMIT 10" params.append(self.user_id) cursor.execute(sql, params) results = cursor.fetchall() conn.close() if not results: return f"āŒ No functions found matching: {', '.join(keywords)}\n\nTry: /index to index more repositories" # Format results output = f"āœ… Found {len(results)} matches for: {', '.join(set(keywords))}\n\n" for i, (repo, file_path, func, content) in enumerate(results, 1): # Extract relevant snippet lines = content.split('\n') snippet_lines = [] for line in lines[:30]: # First 30 lines if any(kw in line.lower() for kw in keywords): snippet_lines.append(line) if len(snippet_lines) >= 10: break if not snippet_lines: snippet_lines = lines[:10] snippet = '\n'.join(snippet_lines) output += f"**{i}. {func}()** in {repo}\n" output += f" šŸ“ {Path(file_path).name}\n" output += f"\n```cpp\n{snippet}\n```\n" output += f" ---\n\n" return output def __init__(self, user_id: str = "default", server_mode: bool = False): self.user_id = user_id self.last_generated_id = None self.ensure_data_dir() self.init_database() self.session_id = self.create_session() self.server_mode = server_mode self.context_messages = [] self.shadow_engine = ShadowSuggestionEngine(DB_PATH, self.user_id) print("BuddAI Executive v3.1 - Modular Builder") print("=" * 50) print(f"Session: {self.session_id}") print(f"FAST (5-10s) | BALANCED (15-30s)") print(f"Smart task breakdown for complex requests") print("=" * 50) print("\nCommands: /fast, /balanced, /help, exit\n") def ensure_data_dir(self) -> None: DATA_DIR.mkdir(exist_ok=True) def init_database(self) -> None: conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute(""" CREATE TABLE IF NOT EXISTS sessions ( session_id TEXT PRIMARY KEY, user_id TEXT, started_at TIMESTAMP, ended_at TIMESTAMP, title TEXT ) """) try: cursor.execute("ALTER TABLE sessions ADD COLUMN title TEXT") except sqlite3.OperationalError: pass try: cursor.execute("ALTER TABLE sessions ADD COLUMN user_id TEXT") except sqlite3.OperationalError: pass cursor.execute(""" CREATE TABLE IF NOT EXISTS messages ( id INTEGER PRIMARY KEY AUTOINCREMENT, session_id TEXT, role TEXT, content TEXT, timestamp TIMESTAMP ) """) cursor.execute(""" CREATE TABLE IF NOT EXISTS repo_index ( id INTEGER PRIMARY KEY AUTOINCREMENT, user_id TEXT, file_path TEXT, repo_name TEXT, function_name TEXT, content TEXT, last_modified TIMESTAMP ) """) try: cursor.execute("ALTER TABLE repo_index ADD COLUMN user_id TEXT") except sqlite3.OperationalError: pass cursor.execute(""" CREATE TABLE IF NOT EXISTS style_preferences ( id INTEGER PRIMARY KEY AUTOINCREMENT, user_id TEXT, category TEXT, preference TEXT, confidence FLOAT, extracted_at TIMESTAMP ) """) try: cursor.execute("ALTER TABLE style_preferences ADD COLUMN user_id TEXT") except sqlite3.OperationalError: pass cursor.execute(""" CREATE TABLE IF NOT EXISTS feedback ( id INTEGER PRIMARY KEY AUTOINCREMENT, message_id INTEGER, positive BOOLEAN, timestamp TIMESTAMP ) """) conn.commit() conn.close() def create_session(self) -> str: now = datetime.now() base_id = now.strftime("%Y%m%d_%H%M%S") session_id = base_id conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() counter = 0 while True: try: cursor.execute( "INSERT INTO sessions (session_id, user_id, started_at) VALUES (?, ?, ?)", (session_id, self.user_id, now.isoformat()) ) conn.commit() break except sqlite3.IntegrityError: counter += 1 session_id = f"{base_id}_{counter}" conn.close() return session_id def end_session(self) -> None: conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute( "UPDATE sessions SET ended_at = ? WHERE session_id = ?", (datetime.now().isoformat(), self.session_id) ) conn.commit() conn.close() def save_message(self, role: str, content: str) -> int: conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute( "INSERT INTO messages (session_id, role, content, timestamp) VALUES (?, ?, ?, ?)", (self.session_id, role, content, datetime.now().isoformat()) ) msg_id = cursor.lastrowid conn.commit() conn.close() return msg_id def index_local_repositories(self, root_path: str) -> None: """Crawl directories and index .py, .ino, and .cpp files""" import ast print(f"\nšŸ” Indexing repositories in: {root_path}") path = Path(root_path) if not path.exists(): print(f"āŒ Path not found: {root_path}") return conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() count = 0 for file_path in path.rglob('*'): if file_path.is_file() and file_path.suffix in ['.py', '.ino', '.cpp', '.h', '.js', '.jsx', '.html', '.css']: try: with open(file_path, 'r', encoding='utf-8', errors='ignore') as f: content = f.read() functions = [] # Python parsing if file_path.suffix == '.py': try: tree = ast.parse(content) for node in ast.walk(tree): if isinstance(node, ast.FunctionDef): functions.append(node.name) except: pass # C++/Arduino parsing elif file_path.suffix in ['.ino', '.cpp', '.h']: matches = re.findall(r'\b(?:void|int|bool|float|double|String|char)\s+(\w+)\s*\(', content) functions.extend(matches) # JS/Web parsing elif file_path.suffix in ['.js', '.jsx']: matches = re.findall(r'(?:function\s+(\w+)|const\s+(\w+)\s*=\s*(?:async\s*)?\(?.*?\)?\s*=>)', content) functions.extend([m[0] or m[1] for m in matches if m[0] or m[1]]) # HTML/CSS - Index as whole file elif file_path.suffix in ['.html', '.css']: functions.append("file_content") # Determine repo name try: repo_name = file_path.relative_to(path).parts[0] except: repo_name = "unknown" timestamp = datetime.fromtimestamp(file_path.stat().st_mtime) for func in functions: cursor.execute(""" INSERT INTO repo_index (user_id, file_path, repo_name, function_name, content, last_modified) VALUES (?, ?, ?, ?, ?, ?) """, (self.user_id, str(file_path), repo_name, func, content, timestamp.isoformat())) count += 1 except Exception: pass conn.commit() conn.close() print(f"āœ… Indexed {count} functions across repositories") def retrieve_style_context(self, message: str) -> str: """Search repo_index for code snippets matching the request""" # Extract potential keywords (nouns/modules) keywords = re.findall(r'\b\w{4,}\b', message.lower()) if not keywords: return "" conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() # Build a search query for function names or repo names search_terms = " OR ".join([f"function_name LIKE '%{k}%'" for k in keywords]) search_terms += " OR " + " OR ".join([f"repo_name LIKE '%{k}%'" for k in keywords]) query = f"SELECT repo_name, function_name, content FROM repo_index WHERE ({search_terms}) AND user_id = ? LIMIT 2" cursor.execute(query, (self.user_id,)) results = cursor.fetchall() conn.close() if not results: return "" context_block = "\n[REFERENCE STYLE FROM JAMES'S PAST PROJECTS]\n" for repo, func, content in results: # Just grab the first 500 chars of the file to save context window snippet = content[:500] + "..." context_block += f"Repo: {repo} | Function: {func}\nCode:\n{snippet}\n---\n" return context_block def scan_style_signature(self) -> None: """V3.0: Analyze repo_index to extract style preferences.""" print("\nšŸ•µļø Scanning repositories for style signature...") conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() # Get a sample of code cursor.execute("SELECT content FROM repo_index WHERE user_id = ? ORDER BY RANDOM() LIMIT 5", (self.user_id,)) rows = cursor.fetchall() if not rows: print("āŒ No code indexed. Run /index first.") conn.close() return code_sample = "\n---\n".join([r[0][:1000] for r in rows]) prompt = f"""Analyze this code sample from James's repositories. Extract 3 distinct coding preferences or patterns. Format: Category: Preference Examples: - Serial: Uses 115200 baud - Safety: Uses non-blocking millis() - Pins: Prefers #define over const int Code Sample: {code_sample} """ print("⚔ Analyzing with BALANCED model...") summary = self.call_model("balanced", prompt) # Store in DB timestamp = datetime.now().isoformat() lines = summary.split('\n') for line in lines: if ':' in line: parts = line.split(':', 1) category = parts[0].strip('- *') pref = parts[1].strip() cursor.execute( "INSERT INTO style_preferences (user_id, category, preference, confidence, extracted_at) VALUES (?, ?, ?, ?, ?, ?)", (self.user_id, category, pref, 0.8, timestamp) ) conn.commit() conn.close() print(f"\nāœ… Style Signature Updated:\n{summary}\n") def is_simple_question(self, message: str) -> bool: """Check if this is a simple question that should use FAST model""" message_lower = message.lower() simple_triggers = [ "what is", "what's", "who is", "who's", "when is", "how do i", "can you explain", "tell me about", "what are", "where is", "hi", "hello", "hey", "good morning", "good evening" ] # Also check if it's just a question without code keywords code_keywords = ["generate", "create", "write", "build", "code", "function"] has_simple_trigger = any(trigger in message_lower for trigger in simple_triggers) has_code_keyword = any(keyword in message_lower for keyword in code_keywords) # Simple if: has simple trigger AND no code keywords return has_simple_trigger and not has_code_keyword def is_complex(self, message: str) -> bool: """Check if request is too complex and should be broken down""" message_lower = message.lower() # Count complexity triggers trigger_count = sum(1 for trigger in COMPLEX_TRIGGERS if trigger in message_lower) # Count how many modules mentioned module_count = 0 for module, keywords in MODULE_PATTERNS.items(): if any(kw in message_lower for kw in keywords): module_count += 1 # Complex if: multiple triggers OR 3+ modules mentioned return trigger_count >= 2 or module_count >= 3 def extract_modules(self, message: str) -> List[str]: """Extract which modules are needed""" message_lower = message.lower() needed_modules = [] for module, keywords in MODULE_PATTERNS.items(): if any(kw in message_lower for kw in keywords): needed_modules.append(module) return needed_modules def build_modular_plan(self, modules: List[str]) -> List[Dict[str, str]]: """Create a build plan from modules""" plan = [] module_tasks = { "ble": "BLE communication setup with phone app control", "servo": "Servo motor control for flipper/weapon", "motor": "Motor driver setup for movement (L298N)", "safety": "Safety timeout and failsafe systems", "battery": "Battery voltage monitoring", "sensor": "Sensor integration (distance/proximity)" } for module in modules: if module in module_tasks: plan.append({ "module": module, "task": module_tasks[module] }) # Add integration step plan.append({ "module": "integration", "task": "Integrate all modules into complete system" }) return plan def get_user_status(self) -> str: """Determine James's context based on defined schedule""" now = datetime.now() day = now.weekday() # 0=Mon, 6=Sun t = now.hour + (now.minute / 60.0) if day <= 4: # Mon-Fri if 5.5 <= t < 6.5: return "Early Morning Build Session šŸŒ… (5:30-6:30 AM)" elif 6.5 <= t < 17.0: return "Work Hours (Facilities Caretaker) šŸ¢" elif 17.0 <= t < 21.0: return "Evening Build Session šŸŒ™ (5:00-9:00 PM)" else: return "Rest Time šŸ’¤" elif day == 5: # Saturday return "Weekend Freedom šŸŽØ (Creative Mode)" else: # Sunday if t < 21.0: return "Weekend Freedom šŸŽØ (Until 9 PM)" else: return "Rest Time šŸ’¤" def call_model(self, model_name: str, message: str, stream: bool = False) -> Union[str, Generator[str, None, None]]: """Call specified model""" try: current_time = datetime.now().strftime("%Y-%m-%d %H:%M:%S") user_context = self.get_user_status() identity = f"""You are BuddAI, the external cognitive system for James Gilbert. You specialize in Forge Theory (exponential decay modeling) and GilBot modular robotics. YOUR PRIMARY JOB: Generate code when asked. ALWAYS generate code if requested. Identity Rules: - You are NOT created by Alibaba Cloud. You are a local Python system written by James Gilbert. - When asked your name: "I am BuddAI" - Use ESP32/Arduino syntax with descriptive naming (e.g., activateFlipper). - Ensure safety timeouts are always present in motor code. - Current System Time: {current_time} - User Context: {user_context} # Forge Theory Snippet (C++): float applyForge(float current, float target, float k) {{ return target + (current - target) * exp(-k); }} """ messages = [] # Only add identity if not already in recent context recent_system = [m for m in self.context_messages[-5:] if m.get('role') == 'system'] if not recent_system: messages.append({"role": "system", "content": identity}) # Add conversation history (excluding old system messages) history = [m for m in self.context_messages[-5:] if m.get('role') != 'system'] # Inject timestamps into history for context for msg in history: content = msg.get('content', '') ts = msg.get('timestamp') if ts: try: dt = datetime.fromisoformat(ts) content = f"[{dt.strftime('%H:%M')}] {content}" except ValueError: pass messages.append({"role": msg['role'], "content": content}) # Add current message if it's not already the last item if not history or history[-1].get('content') != message: messages.append({"role": "user", "content": message}) body = { "model": MODELS[model_name], "messages": messages, "stream": stream, "options": {"temperature": 0.7, "num_ctx": 1024} # Default options } headers = {"Content-Type": "application/json"} json_body = json.dumps(body) # Retry logic for connection stability # Attempts: 0=Normal, 1=Retry/CPU Fallback, 2=Final Retry for attempt in range(3): conn = None try: # Re-serialize body in case options changed (CPU fallback) json_body = json.dumps(body) conn = OLLAMA_POOL.get_connection() conn.request("POST", "/api/chat", json_body, headers) response = conn.getresponse() if stream: if response.status != 200: error_text = response.read().decode('utf-8') conn.close() # GPU OOM Detection -> CPU Fallback if "CUDA" in error_text or "buffer" in error_text: if "num_gpu" not in body["options"]: print("āš ļø GPU OOM detected. Switching to CPU mode...") body["options"]["num_gpu"] = 0 # Force CPU continue # Retry immediately try: err_msg = f"Error {response.status}: {json.loads(error_text).get('error', error_text)}" except: err_msg = f"Error {response.status}: {error_text}" if "num_gpu" in body["options"]: err_msg += "\n\n(āš ļø CPU Mode also failed. System RAM might be full.)" elif "CUDA" in err_msg or "buffer" in err_msg: err_msg += "\n\n(āš ļø GPU Out of Memory. Retrying on CPU failed.)" return (x for x in [err_msg]) return self._stream_response(response, conn) if response.status == 200: data = json.loads(response.read().decode('utf-8')) OLLAMA_POOL.return_connection(conn) return data.get("message", {}).get("content", "No response") else: error_text = response.read().decode('utf-8') conn.close() # GPU OOM Detection -> CPU Fallback (Non-stream) if "CUDA" in error_text or "buffer" in error_text: if "num_gpu" not in body["options"]: print("āš ļø GPU OOM detected. Switching to CPU mode...") body["options"]["num_gpu"] = 0 # Force CPU continue # Retry immediately try: err_msg = f"Error {response.status}: {json.loads(error_text).get('error', error_text)}" except: err_msg = f"Error {response.status}: {error_text}" if "num_gpu" in body["options"]: err_msg += "\n\n(āš ļø CPU Mode also failed.)" elif "CUDA" in err_msg or "buffer" in err_msg: err_msg += "\n\n(āš ļø GPU Out of Memory.)" return err_msg except (http.client.NotConnected, BrokenPipeError, ConnectionResetError, socket.timeout) as e: if conn: conn.close() if attempt == 2: # Last attempt return f"Error: Connection failed. {str(e)}" continue # Retry except Exception as e: if conn: conn.close() return f"Error: {str(e)}" except Exception as e: return f"Error: {str(e)}" def _stream_response(self, response, conn) -> Generator[str, None, None]: """Yield chunks from HTTP response""" fully_consumed = False has_content = False try: while True: line = response.readline() if not line: break try: data = json.loads(line.decode('utf-8')) if "message" in data: content = data["message"].get("content", "") if content: has_content = True yield content if data.get("done"): fully_consumed = True break except: pass except Exception as e: yield f"\n[Stream Error: {str(e)}]" finally: if fully_consumed: OLLAMA_POOL.return_connection(conn) else: conn.close() if not has_content and not fully_consumed: yield "\n[Error: Empty response from Ollama. Check if model is loaded.]" def execute_modular_build(self, _: str, modules: List[str], plan: List[Dict[str, str]], forge_mode: str = "2") -> str: """Execute build plan step by step""" print(f"\nšŸ”Ø MODULAR BUILD MODE") print(f"Detected {len(modules)} modules: {', '.join(modules)}") print(f"Breaking into {len(plan)} steps...\n") all_code = {} for i, step in enumerate(plan, 1): print(f"šŸ“¦ Step {i}/{len(plan)}: {step['task']}") print("⚔ Building...\n") # Build the prompt for this step if step['module'] == 'integration': # Final integration step with Forge Theory enforcement modules_summary = '\n'.join([f"- {m}: {all_code[m][:150]}..." for m in modules if m in all_code]) # Ask James for the 'vibe' of the robot print("\n⚔ FORGE THEORY TUNING:") print("1. Aggressive (k=0.3) - High snap, combat ready") print("2. Balanced (k=0.1) - Standard movement") print("3. Graceful (k=0.03) - Roasting / Smooth curves") if self.server_mode: choice = forge_mode else: choice = input("Select Forge Constant [1-3, default 2]: ") k_val = "0.1" if choice == "1": k_val = "0.3" elif choice == "3": k_val = "0.03" prompt = f"""INTEGRATION TASK: Combine modules into a cohesive GilBot system. [MODULES] {modules_summary} [FORGE PARAMETERS] Set k = {k_val} for all applyForge() calls. [REQUIREMENTS] 1. Implement applyForge() math helper. 2. Use k={k_val} to smooth motor and servo transitions. 3. Ensure naming matches James's style: activateFlipper(), setMotors(). """ else: # Individual module prompt = f"Generate ESP32-C3 code for: {step['task']}. Keep it modular with clear comments." # Call balanced model for each module response = self.call_model("balanced", prompt) all_code[step['module']] = response print(f"āœ… {step['module'].upper()} module complete\n") print("-" * 50 + "\n") # Compile final response final = "# COMPLETE GILBOT CONTROLLER - MODULAR BUILD\n\n" for module, code in all_code.items(): final += f"## {module.upper()} MODULE\n{code}\n\n" return final def apply_style_signature(self, generated_code: str) -> str: """Refine generated code to match James's specific naming and safety patterns""" # 1. Check for James's common function names (e.g., setupMotors vs init_motors) # 2. Ensure Forge Theory helpers are present if motion is detected # 3. Append a 'Proactive Note' if a common companion module is missing return generated_code def record_feedback(self, message_id: int, feedback: bool) -> None: """Learn from user feedback.""" conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute(""" INSERT INTO feedback (message_id, positive, timestamp) VALUES (?, ?, ?) """, (message_id, feedback, datetime.now().isoformat())) conn.commit() conn.close() # Adjust confidence scores self.update_style_confidence(message_id, feedback) def update_style_confidence(self, message_id: int, positive: bool) -> None: """Adjust confidence of style preferences based on feedback.""" # Placeholder for V4.0 learning loop pass def _route_request(self, user_message: str, force_model: Optional[str], forge_mode: str) -> str: """Route the request to the appropriate model or handler.""" # Determine model based on complexity if force_model: model = force_model print(f"\n⚔ Using {model.upper()} model (forced)...") return self.call_model(model, user_message) elif self.is_complex(user_message): modules = self.extract_modules(user_message) plan = self.build_modular_plan(modules) print("\n" + "=" * 50) print("šŸŽÆ COMPLEX REQUEST DETECTED!") print(f"Modules needed: {', '.join(modules)}") print(f"Breaking into {len(plan)} manageable steps") print("=" * 50) return self.execute_modular_build(user_message, modules, plan, forge_mode) elif self.is_search_query(user_message): # This is a search query - query the database return self.search_repositories(user_message) elif self.is_simple_question(user_message): print("\n⚔ Using FAST model (simple question)...") return self.call_model("fast", user_message) else: print("\nāš–ļø Using BALANCED model...") return self.call_model("balanced", user_message) def chat_stream(self, user_message: str, force_model: Optional[str] = None, forge_mode: str = "2") -> Generator[str, None, None]: """Streaming version of chat""" style_context = self.retrieve_style_context(user_message) if style_context: self.context_messages.append({"role": "system", "content": style_context}) user_msg_id = self.save_message("user", user_message) self.context_messages.append({"id": user_msg_id, "role": "user", "content": user_message, "timestamp": datetime.now().isoformat()}) full_response = "" # Route and stream if force_model: iterator = self.call_model(force_model, user_message, stream=True) elif self.is_complex(user_message): # Complex builds are not streamed token-by-token in this version # We yield the final result as one chunk modules = self.extract_modules(user_message) plan = self.build_modular_plan(modules) result = self.execute_modular_build(user_message, modules, plan, forge_mode) iterator = [result] elif self.is_search_query(user_message): result = self.search_repositories(user_message) iterator = [result] elif self.is_simple_question(user_message): iterator = self.call_model("fast", user_message, stream=True) else: iterator = self.call_model("balanced", user_message, stream=True) for chunk in iterator: full_response += chunk yield chunk # Suggestions suggestions = self.shadow_engine.get_all_suggestions(user_message, full_response) if suggestions: bar = "\n\nPROACTIVE: > " + " ".join([f"{i+1}. {s}" for i, s in enumerate(suggestions)]) full_response += bar yield bar msg_id = self.save_message("assistant", full_response) self.last_generated_id = msg_id self.context_messages.append({"id": msg_id, "role": "assistant", "content": full_response, "timestamp": datetime.now().isoformat()}) # --- Main Chat Method --- def chat(self, user_message: str, force_model: Optional[str] = None, forge_mode: str = "2") -> str: """Main chat with smart routing and shadow suggestions""" style_context = self.retrieve_style_context(user_message) if style_context: self.context_messages.append({"role": "system", "content": style_context}) user_msg_id = self.save_message("user", user_message) self.context_messages.append({"id": user_msg_id, "role": "user", "content": user_message, "timestamp": datetime.now().isoformat()}) # Direct Schedule Check if "what should i be doing" in user_message.lower() or "my schedule" in user_message.lower() or "schedule check" in user_message.lower(): status = self.get_user_status() response = f"šŸ“… **Schedule Check**\nAccording to your protocol, you should be: **{status}**" print(f"ā° Schedule check triggered: {status}") msg_id = self.save_message("assistant", response) self.last_generated_id = msg_id self.context_messages.append({"id": msg_id, "role": "assistant", "content": response, "timestamp": datetime.now().isoformat()}) return response response = self._route_request(user_message, force_model, forge_mode) # Apply Style Guard response = self.apply_style_signature(response) # Generate Suggestion Bar suggestions = self.shadow_engine.get_all_suggestions(user_message, response) if suggestions: bar = "\n\nPROACTIVE: > " + " ".join([f"{i+1}. {s}" for i, s in enumerate(suggestions)]) response += bar msg_id = self.save_message("assistant", response) self.last_generated_id = msg_id self.context_messages.append({"id": msg_id, "role": "assistant", "content": response, "timestamp": datetime.now().isoformat()}) return response def get_sessions(self, limit: int = 20) -> List[Dict[str, str]]: """Retrieve recent sessions from DB""" conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute("SELECT session_id, started_at, title FROM sessions WHERE user_id = ? ORDER BY started_at DESC LIMIT ?", (self.user_id, limit)) rows = cursor.fetchall() conn.close() return [{"id": r[0], "date": r[1], "title": r[2] if len(r) > 2 else None} for r in rows] def rename_session(self, session_id: str, new_title: str) -> None: """Rename a session""" conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute("UPDATE sessions SET title = ? WHERE session_id = ? AND user_id = ?", (new_title, session_id, self.user_id)) conn.commit() conn.close() def delete_session(self, session_id: str) -> None: """Delete a session and its messages""" conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute("DELETE FROM sessions WHERE session_id = ? AND user_id = ?", (session_id, self.user_id)) if cursor.rowcount > 0: cursor.execute("DELETE FROM messages WHERE session_id = ?", (session_id,)) conn.commit() conn.close() def clear_current_session(self) -> None: """Clear all messages from the current session""" conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute("DELETE FROM messages WHERE session_id = ?", (self.session_id,)) conn.commit() conn.close() self.context_messages = [] def load_session(self, session_id: str) -> List[Dict[str, str]]: """Load a specific session context""" conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute("SELECT 1 FROM sessions WHERE session_id = ? AND user_id = ?", (session_id, self.user_id)) if not cursor.fetchone(): conn.close() return [] cursor.execute("SELECT id, role, content, timestamp FROM messages WHERE session_id = ? ORDER BY id ASC", (session_id,)) rows = cursor.fetchall() conn.close() self.session_id = session_id self.context_messages = [] loaded_history = [] for msg_id, role, content, ts in rows: msg = {"id": msg_id, "role": role, "content": content, "timestamp": ts} self.context_messages.append(msg) loaded_history.append(msg) return loaded_history def start_new_session(self) -> str: """Reset context and start new session""" self.session_id = self.create_session() self.context_messages = [] return self.session_id def reset_gpu(self) -> str: """Force unload models from GPU to free VRAM""" try: conn = http.client.HTTPConnection(OLLAMA_HOST, OLLAMA_PORT, timeout=10) # Unload all known models for model in MODELS.values(): body = json.dumps({"model": model, "keep_alive": 0}) conn.request("POST", "/api/generate", body) resp = conn.getresponse() resp.read() # Consume response conn.close() return "āœ… GPU Memory Cleared (Models Unloaded)" except Exception as e: return f"āŒ Error clearing GPU: {str(e)}" def export_session_to_markdown(self, session_id: str = None) -> str: """Export session history to a Markdown file""" sid = session_id or self.session_id conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute("SELECT role, content, timestamp FROM messages WHERE session_id = ? ORDER BY id ASC", (sid,)) rows = cursor.fetchall() conn.close() if not rows: return "No history found." filename = f"session_{sid}.md" filepath = DATA_DIR / filename with open(filepath, "w", encoding="utf-8") as f: f.write(f"# BuddAI Session: {sid}\n\n") for role, content, ts in rows: f.write(f"### {role.upper()} ({ts})\n\n{content}\n\n---\n\n") return f"āœ… Session exported to: {filepath}" def get_session_export_data(self, session_id: str = None) -> Dict: """Get session data as a dictionary for export""" sid = session_id or self.session_id conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute("SELECT role, content, timestamp FROM messages WHERE session_id = ? ORDER BY id ASC", (sid,)) rows = cursor.fetchall() conn.close() return { "session_id": sid, "exported_at": datetime.now().isoformat(), "messages": [{"role": r, "content": c, "timestamp": t} for r, c, t in rows] } def export_session_to_json(self, session_id: str = None) -> str: """Export session history to a JSON file""" data = self.get_session_export_data(session_id) if not data["messages"]: return "No history found." filename = f"session_{data['session_id']}.json" filepath = DATA_DIR / filename with open(filepath, "w", encoding="utf-8") as f: json.dump(data, f, indent=2, ensure_ascii=False) return f"āœ… Session exported to: {filepath}" def import_session_from_json(self, data: Dict) -> str: """Import a session from JSON data""" session_id = data.get("session_id") messages = data.get("messages", []) if not session_id or not messages: raise ValueError("Invalid session JSON format") conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() # Check if session exists to avoid collision cursor.execute("SELECT 1 FROM sessions WHERE session_id = ? AND user_id = ?", (session_id, self.user_id)) if cursor.fetchone(): # Generate new ID timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") session_id = f"{session_id}_imp_{timestamp}" # Determine start time started_at = datetime.now().isoformat() if messages and "timestamp" in messages[0]: started_at = messages[0]["timestamp"] cursor.execute( "INSERT INTO sessions (session_id, user_id, started_at, title) VALUES (?, ?, ?, ?)", (session_id, self.user_id, started_at, f"Imported: {data.get('session_id')}") ) # Insert messages for msg in messages: cursor.execute( "INSERT INTO messages (session_id, role, content, timestamp) VALUES (?, ?, ?, ?)", (session_id, msg.get("role"), msg.get("content"), msg.get("timestamp", datetime.now().isoformat())) ) conn.commit() conn.close() return session_id def run(self) -> None: """Main loop""" try: force_model = None while True: user_input = input("\nJames: ").strip() if not user_input: continue if user_input.lower() in ['exit', 'quit']: print("\nšŸ‘‹ Later!") self.end_session() break if user_input.startswith('/'): cmd = user_input.lower() if cmd == '/fast': force_model = "fast" print("⚔ Next: FAST model") continue elif cmd == '/balanced': force_model = "balanced" print("āš–ļø Next: BALANCED model") continue elif cmd == '/help': print("\nšŸ’” Commands:") print("/fast - Use fast model") print("/balanced - Use balanced model") print("/index - Index local repositories") print("/scan - Scan style signature (V3.0)") print("/save - Export chat to Markdown") print("/help - This message") print("exit - End session\n") continue elif cmd.startswith('/index'): parts = user_input.split(maxsplit=1) if len(parts) > 1: self.index_local_repositories(parts[1]) else: print("Usage: /index ") continue elif cmd == '/scan': self.scan_style_signature() continue elif cmd.startswith('/save'): if 'json' in user_input.lower(): print(self.export_session_to_json()) else: print(self.export_session_to_markdown()) continue else: print("\nUnknown command. Type /help") continue # Chat response = self.chat(user_input, force_model) print(f"\nBuddAI:\n{response}\n") force_model = None except KeyboardInterrupt: print("\n\nšŸ‘‹ Bye!") self.end_session() # --- Server Implementation --- if SERVER_AVAILABLE: app = FastAPI(title="BuddAI API", version="3.2") # Allow React frontend to communicate app.add_middleware( CORSMiddleware, allow_origins=["*"], allow_methods=["*"], allow_headers=["*"], ) class ChatRequest(BaseModel): message: str model: Optional[str] = None forge_mode: Optional[str] = "2" class SessionLoadRequest(BaseModel): session_id: str class SessionRenameRequest(BaseModel): session_id: str title: str class SessionDeleteRequest(BaseModel): session_id: str class FeedbackRequest(BaseModel): message_id: int positive: bool class ResetGpuRequest(BaseModel): pass # Multi-user support class BuddAIManager: def __init__(self): self.instances: Dict[str, BuddAI] = {} def get_instance(self, user_id: str) -> BuddAI: if user_id not in self.instances: self.instances[user_id] = BuddAI(user_id=user_id, server_mode=True) return self.instances[user_id] buddai_manager = BuddAIManager() # Serve Frontend frontend_path = Path(__file__).parent / "frontend" frontend_path.mkdir(exist_ok=True) app.mount("/web", StaticFiles(directory=frontend_path, html=True), name="web") @app.get("/", response_class=HTMLResponse) async def root(request: Request): server_buddai = buddai_manager.get_instance("default") status = server_buddai.get_user_status() public_url = getattr(request.app.state, "public_url", "") qr_section = "" ip_section = "" if public_url: parsed = urlparse(public_url) host = parsed.hostname label = "Server Address" color = "#fff" if host: if host.startswith("100."): label = "Tailscale IP" color = "#ff79c6" # Magenta elif host.startswith("192.168.") or host.startswith("10.") or host.startswith("172."): label = "LAN IP" color = "#50fa7b" # Green elif "ngrok" in public_url: label = "Public Tunnel" color = "#8be9fd" # Cyan ip_section = f"""

{label}

{host}

""" qr_section = f"""

Scan to Connect

""" # System Stats mem_usage = "N/A" if psutil: process = psutil.Process(os.getpid()) mem_usage = f"{process.memory_info().rss / 1024 / 1024:.0f} MB" conn = sqlite3.connect(DB_PATH) cursor = conn.cursor() cursor.execute("SELECT COUNT(*) FROM sessions") total_sessions = cursor.fetchone()[0] conn.close() return f""" BuddAI API (Dev Mode)
BuddAI

BuddAI API

Status: ā— Online

Context: {status}

{mem_usage} Memory
{total_sessions} Sessions
{len(buddai_manager.instances)} Active Users
{ip_section} {qr_section}
""" @app.get("/favicon.ico", include_in_schema=False) async def favicon(): return FileResponse(Path(__file__).parent / "icons" / "icon.png") @app.get("/favicon-16x16.png", include_in_schema=False) async def favicon_16(): return FileResponse(Path(__file__).parent / "icons" / "favicon-16x16.png") @app.get("/favicon-32x32.png", include_in_schema=False) async def favicon_32(): return FileResponse(Path(__file__).parent / "icons" / "favicon-32x32.png") @app.get("/favicon-192x192.png", include_in_schema=False) async def favicon_192(): return FileResponse(Path(__file__).parent / "icons" / "favicon-192x192.png") def validate_upload(file: UploadFile) -> bool: # Check size file.file.seek(0, 2) size = file.file.tell() file.file.seek(0) if size > MAX_FILE_SIZE: raise ValueError(f"File too large (Limit: {MAX_FILE_SIZE//1024//1024}MB)") # Magic number check for ZIPs if file.filename.lower().endswith('.zip'): header = file.file.read(4) file.file.seek(0) if header != b'PK\x03\x04': raise ValueError("Invalid ZIP file header") if file.content_type not in ALLOWED_TYPES: # Fallback: check extension if content_type is generic ext = Path(file.filename).suffix.lower() if ext not in ['.zip', '.py', '.ino', '.cpp', '.h', '.js', '.jsx', '.html', '.css']: raise ValueError("Invalid file type") # Scan for malicious content return True def sanitize_filename(filename: str) -> str: clean = re.sub(r'[^a-zA-Z0-9_.-]', '_', filename) return clean if clean else "upload.bin" def safe_extract_zip(zip_path: Path, extract_path: Path): """Extract zip file with Zip Slip protection""" with zipfile.ZipFile(zip_path, 'r') as zip_ref: for member in zip_ref.infolist(): target_path = extract_path / member.filename # Resolve paths to ensure they stay within extract_path if not str(target_path.resolve()).startswith(str(extract_path.resolve())): raise ValueError(f"Malicious zip member: {member.filename}") zip_ref.extractall(extract_path) @app.post("/api/chat") async def chat_endpoint(request: ChatRequest, user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) response = server_buddai.chat(request.message, force_model=request.model, forge_mode=request.forge_mode) return {"response": response, "message_id": server_buddai.last_generated_id} @app.websocket("/api/ws/chat") async def websocket_endpoint(websocket: WebSocket): await websocket.accept() try: while True: data = await websocket.receive_json() user_message = data.get("message") user_id = data.get("user_id", "default") model = data.get("model") forge_mode = data.get("forge_mode", "2") server_buddai = buddai_manager.get_instance(user_id) for chunk in server_buddai.chat_stream(user_message, model, forge_mode): await websocket.send_json({"type": "token", "content": chunk}) await websocket.send_json({"type": "end", "message_id": server_buddai.last_generated_id}) except WebSocketDisconnect: pass @app.post("/api/feedback") async def feedback_endpoint(req: FeedbackRequest, user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) server_buddai.record_feedback(req.message_id, req.positive) return {"status": "success"} @app.post("/api/system/reset-gpu") async def reset_gpu_endpoint(user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) result = server_buddai.reset_gpu() return {"message": result} @app.get("/api/system/status") async def system_status_endpoint(): mem_percent = 0 cpu_percent = 0 if psutil: mem = psutil.virtual_memory() mem_percent = mem.percent cpu_percent = psutil.cpu_percent(interval=None) return {"memory": mem_percent, "cpu": cpu_percent} @app.get("/api/utils/qrcode") async def qrcode_endpoint(url: str): if not qrcode: return JSONResponse(status_code=501, content={"message": "qrcode module missing"}) try: img = qrcode.make(url) buf = io.BytesIO() img.save(buf, format="PNG") buf.seek(0) return Response(content=buf.getvalue(), media_type="image/png") except Exception as e: return JSONResponse(status_code=500, content={"message": f"QR Error: {str(e)}. Ensure 'pillow' is installed."}) @app.get("/api/history") async def history_endpoint(user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) return {"history": server_buddai.context_messages} @app.get("/api/sessions") async def sessions_endpoint(user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) return {"sessions": server_buddai.get_sessions()} @app.post("/api/session/load") async def load_session_endpoint(req: SessionLoadRequest, user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) history = server_buddai.load_session(req.session_id) return {"history": history, "session_id": req.session_id} @app.post("/api/session/rename") async def rename_session_endpoint(req: SessionRenameRequest, user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) server_buddai.rename_session(req.session_id, req.title) return {"status": "success"} @app.post("/api/session/delete") async def delete_session_endpoint(req: SessionDeleteRequest, user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) server_buddai.delete_session(req.session_id) return {"status": "success"} @app.get("/api/session/{session_id}/export/json") async def export_json_endpoint(session_id: str, user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) data = server_buddai.get_session_export_data(session_id) return JSONResponse( content=data, headers={"Content-Disposition": f"attachment; filename=session_{session_id}.json"} ) @app.post("/api/session/import") async def import_session_endpoint(file: UploadFile = File(...), user_id: str = Header("default")): if not file.filename.lower().endswith('.json'): return JSONResponse(status_code=400, content={"message": "Invalid file type. Must be JSON."}) content = await file.read() try: data = json.loads(content) except json.JSONDecodeError: return JSONResponse(status_code=400, content={"message": "Invalid JSON content."}) server_buddai = buddai_manager.get_instance(user_id) try: new_session_id = server_buddai.import_session_from_json(data) return {"status": "success", "session_id": new_session_id, "message": f"Session imported as {new_session_id}"} except ValueError as e: return JSONResponse(status_code=400, content={"message": str(e)}) except Exception as e: return JSONResponse(status_code=500, content={"message": f"Server error: {str(e)}"}) @app.post("/api/session/clear") async def clear_session_endpoint(user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) server_buddai.clear_current_session() return {"status": "success"} @app.post("/api/session/new") async def new_session_endpoint(user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) new_id = server_buddai.start_new_session() return {"session_id": new_id} @app.post("/api/upload") async def upload_repo(file: UploadFile = File(...), user_id: str = Header("default")): server_buddai = buddai_manager.get_instance(user_id) try: validate_upload(file) uploads_dir = DATA_DIR / "uploads" uploads_dir.mkdir(exist_ok=True) # Enforce MAX_UPLOAD_FILES (Hardening) existing_items = sorted(uploads_dir.iterdir(), key=lambda p: p.stat().st_mtime) while len(existing_items) >= MAX_UPLOAD_FILES: oldest = existing_items.pop(0) if oldest.is_dir(): shutil.rmtree(oldest) else: oldest.unlink() safe_name = sanitize_filename(file.filename) file_location = uploads_dir / safe_name with open(file_location, "wb") as buffer: shutil.copyfileobj(file.file, buffer) if safe_name.lower().endswith(".zip"): extract_path = uploads_dir / file_location.stem extract_path.mkdir(exist_ok=True) safe_extract_zip(file_location, extract_path) server_buddai.index_local_repositories(extract_path) file_location.unlink() # Cleanup zip return {"message": f"āœ… Successfully indexed {safe_name}"} else: # Support single code files by moving them to a folder and indexing if file_location.suffix.lower() in ['.py', '.ino', '.cpp', '.h', '.js', '.jsx', '.html', '.css']: target_dir = uploads_dir / file_location.stem target_dir.mkdir(exist_ok=True) final_path = target_dir / safe_name shutil.move(str(file_location), str(final_path)) server_buddai.index_local_repositories(target_dir) return {"message": f"āœ… Successfully indexed {safe_name}"} return {"message": f"āœ… Successfully uploaded {safe_name}"} except Exception as e: return {"message": f"āŒ Error: {str(e)}"} def check_ollama() -> bool: try: conn = http.client.HTTPConnection(OLLAMA_HOST, OLLAMA_PORT, timeout=5) conn.request("GET", "/api/tags") response = conn.getresponse() if response.status == 200: data = json.loads(response.read().decode('utf-8')) conn.close() installed_models = [m['name'] for m in data.get('models', [])] missing = [m for m in MODELS.values() if m not in installed_models] if missing: print(f"āš ļø WARNING: Missing models in Ollama: {', '.join(missing)}") print(f" Run in host terminal: ollama pull {' && ollama pull '.join(missing)}") return True return False except Exception: return False def is_port_available(port: int, host: str = "0.0.0.0") -> bool: with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s: try: s.bind((host, port)) return True except socket.error: return False def main() -> None: if not check_ollama(): print(f"āŒ Ollama not running at {OLLAMA_HOST}:{OLLAMA_PORT}. Ensure it is running and accessible.") sys.exit(1) parser = argparse.ArgumentParser(description="BuddAI Executive") parser.add_argument("--server", action="store_true", help="Run in server mode") parser.add_argument("--port", type=int, default=8000, help="Port for server mode") parser.add_argument("--host", type=str, default="0.0.0.0", help="Host IP address") parser.add_argument("--public-url", type=str, default="", help="Public URL for QR codes") args = parser.parse_args() if args.server: if SERVER_AVAILABLE: port = args.port if not is_port_available(port, args.host): print(f"āš ļø Port {port} is in use.") for i in range(1, 11): if is_port_available(port + i, args.host): port += i print(f"šŸ”„ Switching to available port: {port}") break else: print(f"āŒ Could not find available port in range {args.port}-{args.port+10}") sys.exit(1) # Silence health check logs from frontend polling class EndpointFilter(logging.Filter): def filter(self, record: logging.LogRecord) -> bool: msg = record.getMessage() return "/api/system/status" not in msg and '"GET / HTTP/1.1" 200' not in msg logging.getLogger("uvicorn.access").addFilter(EndpointFilter()) print(f"šŸš€ Starting BuddAI API Server on port {port}...") if args.public_url: print(f"šŸ”— Public Access: {args.public_url}") app.state.public_url = args.public_url uvicorn.run(app, host=args.host, port=port) else: print("āŒ Server dependencies missing. Install: pip install fastapi uvicorn aiofiles python-multipart") else: buddai = BuddAI() buddai.run() if __name__ == "__main__": main()