# PentAGI
Penetration testing Artificial General Intelligence

> **Join the Community!** Connect with security researchers, AI enthusiasts, and fellow ethical hackers. Get support, share insights, and stay updated with the latest PentAGI developments. [![Discord](https://img.shields.io/badge/Discord-7289DA?logo=discord&logoColor=white)](https://discord.gg/2xrMh7qX6m)⠀[![Telegram](https://img.shields.io/badge/Telegram-2CA5E0?logo=telegram&logoColor=white)](https://t.me/+Ka9i6CNwe71hMWQy) vxcontrol%2Fpentagi | Trendshift
## Table of Contents - [Overview](#overview) - [Features](#features) - [Architecture](#architecture) - [Agent Supervision](#advanced-agent-supervision) - [Quick Start](#quick-start) - [API Access](#api-access) - [LLM Provider Configuration](#custom-llm-provider-configuration) - [Ollama](#ollama-provider-configuration) - [OpenAI](#openai-provider-configuration) - [Anthropic](#anthropic-provider-configuration) - [Google AI (Gemini)](#google-ai-gemini-provider-configuration) - [AWS Bedrock](#aws-bedrock-provider-configuration) - [DeepSeek](#deepseek-provider-configuration) - [GLM](#glm-provider-configuration) - [Kimi](#kimi-provider-configuration) - [Qwen](#qwen-provider-configuration) - [Advanced Setup](#advanced-setup) - [Langfuse Integration](#langfuse-integration) - [Monitoring and Observability](#monitoring-and-observability) - [Knowledge Graph (Graphiti)](#knowledge-graph-integration-graphiti) - [OAuth Integration](#github-and-google-oauth-integration) - [Docker Image Configuration](#docker-image-configuration) - [Development](#development) - [Testing LLM Agents](#testing-llm-agents) - [Embedding Configuration and Testing](#embedding-configuration-and-testing) - [Function Testing with ftester](#function-testing-with-ftester) - [Building](#building) - [Credits](#credits) - [License](#license) ## Overview PentAGI is an innovative tool for automated security testing that leverages cutting-edge artificial intelligence technologies. The project is designed for information security professionals, researchers, and enthusiasts who need a powerful and flexible solution for conducting penetration tests. You can watch the video **PentAGI overview**: [![PentAGI Overview Video](https://github.com/user-attachments/assets/0828dc3e-15f1-4a1d-858e-9696a146e478)](https://youtu.be/R70x5Ddzs1o) ## Features - Secure & Isolated. All operations are performed in a sandboxed Docker environment with complete isolation. - Fully Autonomous. AI-powered agent that automatically determines and executes penetration testing steps with optional execution monitoring and intelligent task planning for enhanced reliability. - Professional Pentesting Tools. Built-in suite of 20+ professional security tools including nmap, metasploit, sqlmap, and more. - Smart Memory System. Long-term storage of research results and successful approaches for future use. - Knowledge Graph Integration. Graphiti-powered knowledge graph using Neo4j for semantic relationship tracking and advanced context understanding. - Web Intelligence. Built-in browser via [scraper](https://hub.docker.com/r/vxcontrol/scraper) for gathering latest information from web sources. - External Search Systems. Integration with advanced search APIs including [Tavily](https://tavily.com), [Traversaal](https://traversaal.ai), [Perplexity](https://www.perplexity.ai), [DuckDuckGo](https://duckduckgo.com/), [Google Custom Search](https://programmablesearchengine.google.com/), [Sploitus Search](https://sploitus.com) and [Searxng](https://searxng.org) for comprehensive information gathering. - Team of Specialists. Delegation system with specialized AI agents for research, development, and infrastructure tasks, enhanced with optional execution monitoring and intelligent task planning for optimal performance with smaller models. - Comprehensive Monitoring. Detailed logging and integration with Grafana/Prometheus for real-time system observation. - Detailed Reporting. Generation of thorough vulnerability reports with exploitation guides. - Smart Container Management. Automatic Docker image selection based on specific task requirements. - Modern Interface. Clean and intuitive web UI for system management and monitoring. - Comprehensive APIs. Full-featured REST and GraphQL APIs with Bearer token authentication for automation and integration. - Persistent Storage. All commands and outputs are stored in PostgreSQL with [pgvector](https://hub.docker.com/r/vxcontrol/pgvector) extension. - Scalable Architecture. Microservices-based design supporting horizontal scaling. - Self-Hosted Solution. Complete control over your deployment and data. - Flexible Authentication. Support for 10+ LLM providers ([OpenAI](https://platform.openai.com/), [Anthropic](https://www.anthropic.com/), [Google AI/Gemini](https://ai.google.dev/), [AWS Bedrock](https://aws.amazon.com/bedrock/), [Ollama](https://ollama.com/), [DeepSeek](https://www.deepseek.com/en/), [GLM](https://z.ai/), [Kimi](https://platform.moonshot.ai/), [Qwen](https://www.alibabacloud.com/en/), Custom) plus aggregators ([OpenRouter](https://openrouter.ai/), [DeepInfra](https://deepinfra.com/)). For production local deployments, see our [vLLM + Qwen3.5-27B-FP8 guide](examples/guides/vllm-qwen35-27b-fp8.md). - API Token Authentication. Secure Bearer token system for programmatic access to REST and GraphQL APIs. - Quick Deployment. Easy setup through [Docker Compose](https://docs.docker.com/compose/) with comprehensive environment configuration. ## Architecture ### System Context ```mermaid flowchart TB classDef person fill:#08427B,stroke:#073B6F,color:#fff classDef system fill:#1168BD,stroke:#0B4884,color:#fff classDef external fill:#666666,stroke:#0B4884,color:#fff pentester["👤 Security Engineer (User of the system)"] pentagi["✨ PentAGI (Autonomous penetration testing system)"] target["🎯 target-system (System under test)"] llm["🧠 llm-provider (OpenAI/Anthropic/Ollama/Bedrock/Gemini/Custom)"] search["🔍 search-systems (Google/DuckDuckGo/Tavily/Traversaal/Perplexity/Sploitus/Searxng)"] langfuse["📊 langfuse-ui (LLM Observability Dashboard)"] grafana["📈 grafana (System Monitoring Dashboard)"] pentester --> |Uses HTTPS| pentagi pentester --> |Monitors AI HTTPS| langfuse pentester --> |Monitors System HTTPS| grafana pentagi --> |Tests Various protocols| target pentagi --> |Queries HTTPS| llm pentagi --> |Searches HTTPS| search pentagi --> |Reports HTTPS| langfuse pentagi --> |Reports HTTPS| grafana class pentester person class pentagi system class target,llm,search,langfuse,grafana external linkStyle default stroke:#ffffff,color:#ffffff ```
Container Architecture (click to expand) ```mermaid graph TB subgraph Core Services UI[Frontend UI
React + TypeScript] API[Backend API
Go + GraphQL] DB[(Vector Store
PostgreSQL + pgvector)] MQ[Task Queue
Async Processing] Agent[AI Agents
Multi-Agent System] end subgraph Knowledge Graph Graphiti[Graphiti
Knowledge Graph API] Neo4j[(Neo4j
Graph Database)] end subgraph Monitoring Grafana[Grafana
Dashboards] VictoriaMetrics[VictoriaMetrics
Time-series DB] Jaeger[Jaeger
Distributed Tracing] Loki[Loki
Log Aggregation] OTEL[OpenTelemetry
Data Collection] end subgraph Analytics Langfuse[Langfuse
LLM Analytics] ClickHouse[ClickHouse
Analytics DB] Redis[Redis
Cache + Rate Limiter] MinIO[MinIO
S3 Storage] end subgraph Security Tools Scraper[Web Scraper
Isolated Browser] PenTest[Security Tools
20+ Pro Tools
Sandboxed Execution] end UI --> |HTTP/WS| API API --> |SQL| DB API --> |Events| MQ MQ --> |Tasks| Agent Agent --> |Commands| PenTest Agent --> |Queries| DB Agent --> |Knowledge| Graphiti Graphiti --> |Graph| Neo4j API --> |Telemetry| OTEL OTEL --> |Metrics| VictoriaMetrics OTEL --> |Traces| Jaeger OTEL --> |Logs| Loki Grafana --> |Query| VictoriaMetrics Grafana --> |Query| Jaeger Grafana --> |Query| Loki API --> |Analytics| Langfuse Langfuse --> |Store| ClickHouse Langfuse --> |Cache| Redis Langfuse --> |Files| MinIO classDef core fill:#f9f,stroke:#333,stroke-width:2px,color:#000 classDef knowledge fill:#ffa,stroke:#333,stroke-width:2px,color:#000 classDef monitoring fill:#bbf,stroke:#333,stroke-width:2px,color:#000 classDef analytics fill:#bfb,stroke:#333,stroke-width:2px,color:#000 classDef tools fill:#fbb,stroke:#333,stroke-width:2px,color:#000 class UI,API,DB,MQ,Agent core class Graphiti,Neo4j knowledge class Grafana,VictoriaMetrics,Jaeger,Loki,OTEL monitoring class Langfuse,ClickHouse,Redis,MinIO analytics class Scraper,PenTest tools ```
Entity Relationship (click to expand) ```mermaid erDiagram Flow ||--o{ Task : contains Task ||--o{ SubTask : contains SubTask ||--o{ Action : contains Action ||--o{ Artifact : produces Action ||--o{ Memory : stores Flow { string id PK string name "Flow name" string description "Flow description" string status "active/completed/failed" json parameters "Flow parameters" timestamp created_at timestamp updated_at } Task { string id PK string flow_id FK string name "Task name" string description "Task description" string status "pending/running/done/failed" json result "Task results" timestamp created_at timestamp updated_at } SubTask { string id PK string task_id FK string name "Subtask name" string description "Subtask description" string status "queued/running/completed/failed" string agent_type "researcher/developer/executor" json context "Agent context" timestamp created_at timestamp updated_at } Action { string id PK string subtask_id FK string type "command/search/analyze/etc" string status "success/failure" json parameters "Action parameters" json result "Action results" timestamp created_at } Artifact { string id PK string action_id FK string type "file/report/log" string path "Storage path" json metadata "Additional info" timestamp created_at } Memory { string id PK string action_id FK string type "observation/conclusion" vector embedding "Vector representation" text content "Memory content" timestamp created_at } ```
Agent Interaction (click to expand) ```mermaid sequenceDiagram participant O as Orchestrator participant R as Researcher participant D as Developer participant E as Executor participant VS as Vector Store participant KB as Knowledge Base Note over O,KB: Flow Initialization O->>VS: Query similar tasks VS-->>O: Return experiences O->>KB: Load relevant knowledge KB-->>O: Return context Note over O,R: Research Phase O->>R: Analyze target R->>VS: Search similar cases VS-->>R: Return patterns R->>KB: Query vulnerabilities KB-->>R: Return known issues R->>VS: Store findings R-->>O: Research results Note over O,D: Planning Phase O->>D: Plan attack D->>VS: Query exploits VS-->>D: Return techniques D->>KB: Load tools info KB-->>D: Return capabilities D-->>O: Attack plan Note over O,E: Execution Phase O->>E: Execute plan E->>KB: Load tool guides KB-->>E: Return procedures E->>VS: Store results E-->>O: Execution status ```
Memory System (click to expand) ```mermaid graph TB subgraph "Long-term Memory" VS[(Vector Store
Embeddings DB)] KB[Knowledge Base
Domain Expertise] Tools[Tools Knowledge
Usage Patterns] end subgraph "Working Memory" Context[Current Context
Task State] Goals[Active Goals
Objectives] State[System State
Resources] end subgraph "Episodic Memory" Actions[Past Actions
Commands History] Results[Action Results
Outcomes] Patterns[Success Patterns
Best Practices] end Context --> |Query| VS VS --> |Retrieve| Context Goals --> |Consult| KB KB --> |Guide| Goals State --> |Record| Actions Actions --> |Learn| Patterns Patterns --> |Store| VS Tools --> |Inform| State Results --> |Update| Tools VS --> |Enhance| KB KB --> |Index| VS classDef ltm fill:#f9f,stroke:#333,stroke-width:2px,color:#000 classDef wm fill:#bbf,stroke:#333,stroke-width:2px,color:#000 classDef em fill:#bfb,stroke:#333,stroke-width:2px,color:#000 class VS,KB,Tools ltm class Context,Goals,State wm class Actions,Results,Patterns em ```
Chain Summarization (click to expand) The chain summarization system manages conversation context growth by selectively summarizing older messages. This is critical for preventing token limits from being exceeded while maintaining conversation coherence. ```mermaid flowchart TD A[Input Chain] --> B{Needs Summarization?} B -->|No| C[Return Original Chain] B -->|Yes| D[Convert to ChainAST] D --> E[Apply Section Summarization] E --> F[Process Oversized Pairs] F --> G[Manage Last Section Size] G --> H[Apply QA Summarization] H --> I[Rebuild Chain with Summaries] I --> J{Is New Chain Smaller?} J -->|Yes| K[Return Optimized Chain] J -->|No| C classDef process fill:#bbf,stroke:#333,stroke-width:2px,color:#000 classDef decision fill:#bfb,stroke:#333,stroke-width:2px,color:#000 classDef output fill:#fbb,stroke:#333,stroke-width:2px,color:#000 class A,D,E,F,G,H,I process class B,J decision class C,K output ``` The algorithm operates on a structured representation of conversation chains (ChainAST) that preserves message types including tool calls and their responses. All summarization operations maintain critical conversation flow while reducing context size. ### Global Summarizer Configuration Options | Parameter | Environment Variable | Default | Description | | --------------------- | -------------------------------- | ------- | ---------------------------------------------------------- | | Preserve Last | `SUMMARIZER_PRESERVE_LAST` | `true` | Whether to keep all messages in the last section intact | | Use QA Pairs | `SUMMARIZER_USE_QA` | `true` | Whether to use QA pair summarization strategy | | Summarize Human in QA | `SUMMARIZER_SUM_MSG_HUMAN_IN_QA` | `false` | Whether to summarize human messages in QA pairs | | Last Section Size | `SUMMARIZER_LAST_SEC_BYTES` | `51200` | Maximum byte size for last section (50KB) | | Max Body Pair Size | `SUMMARIZER_MAX_BP_BYTES` | `16384` | Maximum byte size for a single body pair (16KB) | | Max QA Sections | `SUMMARIZER_MAX_QA_SECTIONS` | `10` | Maximum QA pair sections to preserve | | Max QA Size | `SUMMARIZER_MAX_QA_BYTES` | `65536` | Maximum byte size for QA pair sections (64KB) | | Keep QA Sections | `SUMMARIZER_KEEP_QA_SECTIONS` | `1` | Number of recent QA sections to keep without summarization | ### Assistant Summarizer Configuration Options Assistant instances can use customized summarization settings to fine-tune context management behavior: | Parameter | Environment Variable | Default | Description | | ------------------ | --------------------------------------- | ------- | -------------------------------------------------------------------- | | Preserve Last | `ASSISTANT_SUMMARIZER_PRESERVE_LAST` | `true` | Whether to preserve all messages in the assistant's last section | | Last Section Size | `ASSISTANT_SUMMARIZER_LAST_SEC_BYTES` | `76800` | Maximum byte size for assistant's last section (75KB) | | Max Body Pair Size | `ASSISTANT_SUMMARIZER_MAX_BP_BYTES` | `16384` | Maximum byte size for a single body pair in assistant context (16KB) | | Max QA Sections | `ASSISTANT_SUMMARIZER_MAX_QA_SECTIONS` | `7` | Maximum QA sections to preserve in assistant context | | Max QA Size | `ASSISTANT_SUMMARIZER_MAX_QA_BYTES` | `76800` | Maximum byte size for assistant's QA sections (75KB) | | Keep QA Sections | `ASSISTANT_SUMMARIZER_KEEP_QA_SECTIONS` | `3` | Number of recent QA sections to preserve without summarization | The assistant summarizer configuration provides more memory for context retention compared to the global settings, preserving more recent conversation history while still ensuring efficient token usage. ### Summarizer Environment Configuration ```bash # Default values for global summarizer logic SUMMARIZER_PRESERVE_LAST=true SUMMARIZER_USE_QA=true SUMMARIZER_SUM_MSG_HUMAN_IN_QA=false SUMMARIZER_LAST_SEC_BYTES=51200 SUMMARIZER_MAX_BP_BYTES=16384 SUMMARIZER_MAX_QA_SECTIONS=10 SUMMARIZER_MAX_QA_BYTES=65536 SUMMARIZER_KEEP_QA_SECTIONS=1 # Default values for assistant summarizer logic ASSISTANT_SUMMARIZER_PRESERVE_LAST=true ASSISTANT_SUMMARIZER_LAST_SEC_BYTES=76800 ASSISTANT_SUMMARIZER_MAX_BP_BYTES=16384 ASSISTANT_SUMMARIZER_MAX_QA_SECTIONS=7 ASSISTANT_SUMMARIZER_MAX_QA_BYTES=76800 ASSISTANT_SUMMARIZER_KEEP_QA_SECTIONS=3 ```
Advanced Agent Supervision (click to expand) PentAGI includes sophisticated multi-layered agent supervision mechanisms to ensure efficient task execution, prevent infinite loops, and provide intelligent recovery from stuck states: ### Execution Monitoring (Beta) - **Automatic Mentor Intervention**: Adviser agent (mentor) is automatically invoked when execution patterns indicate potential issues - **Pattern Detection**: Monitors identical tool calls (threshold: 5, configurable) and total tool calls (threshold: 10, configurable) - **Progress Analysis**: Evaluates whether agent advances toward subtask objective, detects loops and inefficiencies - **Alternative Strategies**: Recommends different approaches when current strategy fails - **Information Retrieval Guidance**: Suggests searching for established solutions instead of reinventing - **Enhanced Response Format**: Tool responses include both `` and `` sections - **Configurable**: Enable via `EXECUTION_MONITOR_ENABLED` (default: false), customize thresholds with `EXECUTION_MONITOR_SAME_TOOL_LIMIT` and `EXECUTION_MONITOR_TOTAL_TOOL_LIMIT` **Best for**: Smaller models (< 32B parameters), complex attack scenarios requiring continuous guidance, preventing agents from getting stuck on single approach **Performance Impact**: 2-3x increase in execution time and token usage, but delivers **2x improvement in result quality** based on testing with Qwen3.5-27B-FP8 ### Intelligent Task Planning (Beta) - **Automated Decomposition**: Planner (adviser in planning mode) generates 3-7 specific, actionable steps before specialist agents begin work - **Context-Aware Plans**: Analyzes full execution context via enricher agent to create informed plans - **Structured Assignment**: Original request wrapped in `` structure with execution plan and instructions - **Scope Management**: Prevents scope creep by keeping agents focused on current subtask only - **Enriched Instructions**: Plans highlight critical actions, potential pitfalls, and verification points - **Configurable**: Enable via `AGENT_PLANNING_STEP_ENABLED` (default: false) **Best for**: Models < 32B parameters, complex penetration testing workflows, improving success rates on sophisticated tasks **Enhanced Adviser Configuration**: Works exceptionally well when adviser agent uses stronger model or enhanced settings. Example: using same base model with maximum reasoning mode for adviser (see [`vllm-qwen3.5-27b-fp8.provider.yml`](examples/configs/vllm-qwen3.5-27b-fp8.provider.yml)) enables comprehensive task analysis and strategic planning from identical model architecture. **Performance Impact**: Adds planning overhead but significantly improves completion rates and reduces redundant work ### Tool Call Limits (Always Active) - **Hard Limits**: Prevent runaway executions regardless of supervision mode status - **Differentiated by Agent Type**: - General agents (Assistant, Primary Agent, Pentester, Coder, Installer): `MAX_GENERAL_AGENT_TOOL_CALLS` (default: 100) - Limited agents (Searcher, Enricher, Memorist, Generator, Reporter, Adviser, Reflector, Planner): `MAX_LIMITED_AGENT_TOOL_CALLS` (default: 20) - **Graceful Termination**: Reflector guides agents to proper completion when approaching limits - **Resource Protection**: Ensures system stability and prevents resource exhaustion ### Reflector Integration (Always Active) - **Automatic Correction**: Invoked when LLM fails to generate tool calls after 3 attempts - **Strategic Guidance**: Analyzes failures and guides agents toward proper tool usage or barrier tools (`done`, `ask`) - **Recovery Mechanism**: Provides contextual guidance based on specific failure patterns - **Limit Enforcement**: Coordinates graceful termination when tool call limits are reached ### Recommendations for Open Source Models **Must-Have for Models < 32B Parameters**: Testing with Qwen3.5-27B-FP8 demonstrates that enabling both Execution Monitoring and Task Planning is **essential** for smaller open source models: - **Quality Improvement**: 2x better results compared to baseline execution without supervision - **Loop Prevention**: Significantly reduces infinite loops and redundant work - **Attack Diversity**: Encourages exploration of multiple attack vectors instead of fixating on single approach - **Air-Gapped Deployments**: Enables production-grade autonomous pentesting in closed network environments with local LLM inference **Trade-offs**: - Token consumption: 2-3x increase due to mentor/planner invocations - Execution time: 2-3x longer due to analysis and planning steps - Result quality: 2x improvement in completeness, accuracy, and attack coverage - Model requirements: Works best when adviser uses enhanced configuration (higher reasoning parameters, stronger model variant, or different model) **Configuration Strategy**: For optimal performance with smaller models, configure adviser agent with enhanced settings: - Use same model with maximum reasoning mode (example: [`vllm-qwen3.5-27b-fp8.provider.yml`](examples/configs/vllm-qwen3.5-27b-fp8.provider.yml)) - Or use stronger model for adviser while keeping base model for other agents - Adjust monitoring thresholds based on task complexity and model capabilities
The architecture of PentAGI is designed to be modular, scalable, and secure. Here are the key components: 1. **Core Services** - Frontend UI: React-based web interface with TypeScript for type safety - Backend API: Go-based REST and GraphQL APIs with Bearer token authentication for programmatic access - Vector Store: PostgreSQL with pgvector for semantic search and memory storage - Task Queue: Async task processing system for reliable operation - AI Agent: Multi-agent system with specialized roles for efficient testing 2. **Knowledge Graph** - Graphiti: Knowledge graph API for semantic relationship tracking and contextual understanding - Neo4j: Graph database for storing and querying relationships between entities, actions, and outcomes - Automatic capturing of agent responses and tool executions for building comprehensive knowledge base 3. **Monitoring Stack** - OpenTelemetry: Unified observability data collection and correlation - Grafana: Real-time visualization and alerting dashboards - VictoriaMetrics: High-performance time-series metrics storage - Jaeger: End-to-end distributed tracing for debugging - Loki: Scalable log aggregation and analysis 4. **Analytics Platform** - Langfuse: Advanced LLM observability and performance analytics - ClickHouse: Column-oriented analytics data warehouse - Redis: High-speed caching and rate limiting - MinIO: S3-compatible object storage for artifacts 5. **Security Tools** - Web Scraper: Isolated browser environment for safe web interaction - Pentesting Tools: Comprehensive suite of 20+ professional security tools - Sandboxed Execution: All operations run in isolated containers 6. **Memory Systems** - Long-term Memory: Persistent storage of knowledge and experiences - Working Memory: Active context and goals for current operations - Episodic Memory: Historical actions and success patterns - Knowledge Base: Structured domain expertise and tool capabilities - Context Management: Intelligently manages growing LLM context windows using chain summarization The system uses Docker containers for isolation and easy deployment, with separate networks for core services, monitoring, and analytics to ensure proper security boundaries. Each component is designed to scale horizontally and can be configured for high availability in production environments. ## Quick Start ### System Requirements - Docker and Docker Compose (or Podman - see [Podman configuration](#running-pentagi-with-podman)) - Minimum 2 vCPU - Minimum 4GB RAM - 20GB free disk space - Internet access for downloading images and updates ### Using Installer (Recommended) PentAGI provides an interactive installer with a terminal-based UI for streamlined configuration and deployment. The installer guides you through system checks, LLM provider setup, search engine configuration, and security hardening. **Supported Platforms:** - **Linux**: amd64 [download](https://pentagi.com/downloads/linux/amd64/installer-latest.zip) | arm64 [download](https://pentagi.com/downloads/linux/arm64/installer-latest.zip) - **Windows**: amd64 [download](https://pentagi.com/downloads/windows/amd64/installer-latest.zip) - **macOS**: amd64 (Intel) [download](https://pentagi.com/downloads/darwin/amd64/installer-latest.zip) | arm64 (M-series) [download](https://pentagi.com/downloads/darwin/arm64/installer-latest.zip) **Quick Installation (Linux amd64):** ```bash # Create installation directory mkdir -p pentagi && cd pentagi # Download installer wget -O installer.zip https://pentagi.com/downloads/linux/amd64/installer-latest.zip # Extract unzip installer.zip # Run interactive installer ./installer ``` **Prerequisites & Permissions:** The installer requires appropriate privileges to interact with the Docker API for proper operation. By default, it uses the Docker socket (`/var/run/docker.sock`) which requires either: - **Option 1 (Recommended for production):** Run the installer as root: ```bash sudo ./installer ``` - **Option 2 (Development environments):** Grant your user access to the Docker socket by adding them to the `docker` group: ```bash # Add your user to the docker group sudo usermod -aG docker $USER # Log out and log back in, or activate the group immediately newgrp docker # Verify Docker access (should run without sudo) docker ps ``` ⚠️ **Security Note:** Adding a user to the `docker` group grants root-equivalent privileges. Only do this for trusted users in controlled environments. For production deployments, consider using rootless Docker mode or running the installer with sudo. The installer will: 1. **System Checks**: Verify Docker, network connectivity, and system requirements 2. **Environment Setup**: Create and configure `.env` file with optimal defaults 3. **Provider Configuration**: Set up LLM providers (OpenAI, Anthropic, Gemini, Bedrock, Ollama, Custom) 4. **Search Engines**: Configure DuckDuckGo, Google, Tavily, Traversaal, Perplexity, Sploitus, Searxng 5. **Security Hardening**: Generate secure credentials and configure SSL certificates 6. **Deployment**: Start PentAGI with docker-compose **For Production & Enhanced Security:** For production deployments or security-sensitive environments, we **strongly recommend** using a distributed two-node architecture where worker operations are isolated on a separate server. This prevents untrusted code execution and network access issues on your main system. **See detailed guide**: [Worker Node Setup](examples/guides/worker_node.md) The two-node setup provides: - **Isolated Execution**: Worker containers run on dedicated hardware - **Network Isolation**: Separate network boundaries for penetration testing - **Security Boundaries**: Docker-in-Docker with TLS authentication - **OOB Attack Support**: Dedicated port ranges for out-of-band techniques ### Manual Installation 1. Create a working directory or clone the repository: ```bash mkdir pentagi && cd pentagi ``` 2. Copy `.env.example` to `.env` or download it: ```bash curl -o .env https://raw.githubusercontent.com/vxcontrol/pentagi/master/.env.example ``` 3. Touch examples files (`example.custom.provider.yml`, `example.ollama.provider.yml`) or download it: ```bash curl -o example.custom.provider.yml https://raw.githubusercontent.com/vxcontrol/pentagi/master/examples/configs/custom-openai.provider.yml curl -o example.ollama.provider.yml https://raw.githubusercontent.com/vxcontrol/pentagi/master/examples/configs/ollama-llama318b.provider.yml ``` 4. Fill in the required API keys in `.env` file. ```bash # Required: At least one of these LLM providers OPEN_AI_KEY=your_openai_key ANTHROPIC_API_KEY=your_anthropic_key GEMINI_API_KEY=your_gemini_key # Optional: AWS Bedrock provider (enterprise-grade models) BEDROCK_REGION=us-east-1 # Choose one authentication method: BEDROCK_DEFAULT_AUTH=true # Option 1: Use AWS SDK default credential chain (recommended for EC2/ECS) # BEDROCK_BEARER_TOKEN=your_bearer_token # Option 2: Bearer token authentication # BEDROCK_ACCESS_KEY_ID=your_aws_access_key # Option 3: Static credentials # BEDROCK_SECRET_ACCESS_KEY=your_aws_secret_key # Optional: Ollama provider (local or cloud) # OLLAMA_SERVER_URL=http://ollama-server:11434 # Local server # OLLAMA_SERVER_URL=https://ollama.com # Cloud service # OLLAMA_SERVER_API_KEY=your_ollama_cloud_key # Required for cloud, empty for local # Optional: Chinese AI providers # DEEPSEEK_API_KEY=your_deepseek_key # DeepSeek (strong reasoning) # GLM_API_KEY=your_glm_key # GLM (Zhipu AI) # KIMI_API_KEY=your_kimi_key # Kimi (Moonshot AI, ultra-long context) # QWEN_API_KEY=your_qwen_key # Qwen (Alibaba Cloud, multimodal) # Optional: Local LLM provider (zero-cost inference) OLLAMA_SERVER_URL=http://localhost:11434 OLLAMA_SERVER_MODEL=your_model_name # Optional: Additional search capabilities DUCKDUCKGO_ENABLED=true DUCKDUCKGO_REGION=us-en DUCKDUCKGO_SAFESEARCH= DUCKDUCKGO_TIME_RANGE= SPLOITUS_ENABLED=true GOOGLE_API_KEY=your_google_key GOOGLE_CX_KEY=your_google_cx TAVILY_API_KEY=your_tavily_key TRAVERSAAL_API_KEY=your_traversaal_key PERPLEXITY_API_KEY=your_perplexity_key PERPLEXITY_MODEL=sonar-pro PERPLEXITY_CONTEXT_SIZE=medium # Searxng meta search engine (aggregates results from multiple sources) SEARXNG_URL=http://your-searxng-instance:8080 SEARXNG_CATEGORIES=general SEARXNG_LANGUAGE= SEARXNG_SAFESEARCH=0 SEARXNG_TIME_RANGE= SEARXNG_TIMEOUT= ## Graphiti knowledge graph settings GRAPHITI_ENABLED=true GRAPHITI_TIMEOUT=30 GRAPHITI_URL=http://graphiti:8000 GRAPHITI_MODEL_NAME=gpt-5-mini # Neo4j settings (used by Graphiti stack) NEO4J_USER=neo4j NEO4J_DATABASE=neo4j NEO4J_PASSWORD=devpassword NEO4J_URI=bolt://neo4j:7687 # Assistant configuration ASSISTANT_USE_AGENTS=false # Default value for agent usage when creating new assistants ``` 5. Change all security related environment variables in `.env` file to improve security.
Security related environment variables ### Main Security Settings - `COOKIE_SIGNING_SALT` - Salt for cookie signing, change to random value - `PUBLIC_URL` - Public URL of your server (eg. `https://pentagi.example.com`) - `SERVER_SSL_CRT` and `SERVER_SSL_KEY` - Custom paths to your existing SSL certificate and key for HTTPS (these paths should be used in the docker-compose.yml file to mount as volumes) ### Scraper Access - `SCRAPER_PUBLIC_URL` - Public URL for scraper if you want to use different scraper server for public URLs - `SCRAPER_PRIVATE_URL` - Private URL for scraper (local scraper server in docker-compose.yml file to access it to local URLs) ### Access Credentials - `PENTAGI_POSTGRES_USER` and `PENTAGI_POSTGRES_PASSWORD` - PostgreSQL credentials - `NEO4J_USER` and `NEO4J_PASSWORD` - Neo4j credentials (for Graphiti knowledge graph)
6. Remove all inline comments from `.env` file if you want to use it in VSCode or other IDEs as a envFile option: ```bash perl -i -pe 's/\s+#.*$//' .env ``` 7. Run the PentAGI stack: ```bash curl -O https://raw.githubusercontent.com/vxcontrol/pentagi/master/docker-compose.yml docker compose up -d ``` Visit [localhost:8443](https://localhost:8443) to access PentAGI Web UI (default is `admin@pentagi.com` / `admin`) > [!NOTE] > If you caught an error about `pentagi-network` or `observability-network` or `langfuse-network` you need to run `docker-compose.yml` firstly to create these networks and after that run `docker-compose-langfuse.yml`, `docker-compose-graphiti.yml`, and `docker-compose-observability.yml` to use Langfuse, Graphiti, and Observability services. > > You have to set at least one Language Model provider (OpenAI, Anthropic, Gemini, AWS Bedrock, or Ollama) to use PentAGI. AWS Bedrock provides enterprise-grade access to multiple foundation models from leading AI companies, while Ollama provides zero-cost local inference if you have sufficient computational resources. Additional API keys for search engines are optional but recommended for better results. > > **For fully local deployment with advanced models**: See our comprehensive guide on [Running PentAGI with vLLM and Qwen3.5-27B-FP8](examples/guides/vllm-qwen35-27b-fp8.md) for a production-grade local LLM setup. This configuration achieves ~13,000 TPS for prompt processing and ~650 TPS for completion on 4× RTX 5090 GPUs, supporting 12+ concurrent flows with complete independence from cloud providers. > > `LLM_SERVER_*` environment variables are experimental feature and will be changed in the future. Right now you can use them to specify custom LLM server URL and one model for all agent types. > > `PROXY_URL` is a global proxy URL for all LLM providers and external search systems. You can use it for isolation from external networks. > > The `docker-compose.yml` file runs the PentAGI service as root user because it needs access to docker.sock for container management. If you're using TCP/IP network connection to Docker instead of socket file, you can remove root privileges and use the default `pentagi` user for better security. ### Accessing PentAGI from External Networks By default, PentAGI binds to `127.0.0.1` (localhost only) for security. To access PentAGI from other machines on your network, you need to configure external access. #### Configuration Steps 1. **Update `.env` file** with your server's IP address: ```bash # Network binding - allow external connections PENTAGI_LISTEN_IP=0.0.0.0 PENTAGI_LISTEN_PORT=8443 # Public URL - use your actual server IP or hostname # Replace 192.168.1.100 with your server's IP address PUBLIC_URL=https://192.168.1.100:8443 # CORS origins - list all URLs that will access PentAGI # Include localhost for local access AND your server IP for external access CORS_ORIGINS=https://localhost:8443,https://192.168.1.100:8443 ``` > [!IMPORTANT] > - Replace `192.168.1.100` with your actual server's IP address > - Do NOT use `0.0.0.0` in `PUBLIC_URL` or `CORS_ORIGINS` - use the actual IP address > - Include both localhost and your server IP in `CORS_ORIGINS` for flexibility 2. **Recreate containers** to apply the changes: ```bash docker compose down docker compose up -d --force-recreate ``` 3. **Verify port binding:** ```bash docker ps | grep pentagi ``` You should see `0.0.0.0:8443->8443/tcp` or `:::8443->8443/tcp`. If you see `127.0.0.1:8443->8443/tcp`, the environment variable wasn't picked up. In this case, directly edit `docker-compose.yml` line 31: ```yaml ports: - "0.0.0.0:8443:8443" ``` Then recreate containers again. 4. **Configure firewall** to allow incoming connections on port 8443: ```bash # Ubuntu/Debian with UFW sudo ufw allow 8443/tcp sudo ufw reload # CentOS/RHEL with firewalld sudo firewall-cmd --permanent --add-port=8443/tcp sudo firewall-cmd --reload ``` 5. **Access PentAGI:** - **Local access:** `https://localhost:8443` - **Network access:** `https://your-server-ip:8443` > [!NOTE] > You'll need to accept the self-signed SSL certificate warning in your browser when accessing via IP address. --- ### Running PentAGI with Podman PentAGI fully supports Podman as a Docker alternative. However, when using **Podman in rootless mode**, the scraper service requires special configuration because rootless containers cannot bind privileged ports (ports below 1024). #### Podman Rootless Configuration The default scraper configuration uses port 443 (HTTPS), which is a privileged port. For Podman rootless, reconfigure the scraper to use a non-privileged port: **1. Edit `docker-compose.yml`** - modify the `scraper` service (around line 199): ```yaml scraper: image: vxcontrol/scraper:latest restart: unless-stopped container_name: scraper hostname: scraper expose: - 3000/tcp # Changed from 443 to 3000 ports: - "${SCRAPER_LISTEN_IP:-127.0.0.1}:${SCRAPER_LISTEN_PORT:-9443}:3000" # Map to port 3000 environment: - MAX_CONCURRENT_SESSIONS=${LOCAL_SCRAPER_MAX_CONCURRENT_SESSIONS:-10} - USERNAME=${LOCAL_SCRAPER_USERNAME:-someuser} - PASSWORD=${LOCAL_SCRAPER_PASSWORD:-somepass} logging: options: max-size: 50m max-file: "7" volumes: - scraper-ssl:/usr/src/app/ssl networks: - pentagi-network shm_size: 2g ``` **2. Update `.env` file** - change the scraper URL to use HTTP and port 3000: ```bash # Scraper configuration for Podman rootless SCRAPER_PRIVATE_URL=http://someuser:somepass@scraper:3000/ LOCAL_SCRAPER_USERNAME=someuser LOCAL_SCRAPER_PASSWORD=somepass ``` > [!IMPORTANT] > Key changes for Podman: > - Use **HTTP** instead of HTTPS for `SCRAPER_PRIVATE_URL` > - Use port **3000** instead of 443 > - Change internal `expose` to `3000/tcp` > - Update port mapping to target `3000` instead of `443` **3. Recreate containers:** ```bash podman-compose down podman-compose up -d --force-recreate ``` **4. Test scraper connectivity:** ```bash # Test from within the pentagi container podman exec -it pentagi wget -O- "http://someuser:somepass@scraper:3000/html?url=http://example.com" ``` If you see HTML output, the scraper is working correctly. #### Podman Rootful Mode If you're running Podman in rootful mode (with sudo), you can use the default configuration without modifications. The scraper will work on port 443 as intended. #### Docker Compatibility All Podman configurations remain fully compatible with Docker. The non-privileged port approach works identically on both container runtimes. ### Assistant Configuration PentAGI allows you to configure default behavior for assistants: | Variable | Default | Description | | ---------------------- | ------- | ----------------------------------------------------------------------- | | `ASSISTANT_USE_AGENTS` | `false` | Controls the default value for agent usage when creating new assistants | The `ASSISTANT_USE_AGENTS` setting affects the initial state of the "Use Agents" toggle when creating a new assistant in the UI: - `false` (default): New assistants are created with agent delegation disabled by default - `true`: New assistants are created with agent delegation enabled by default Note that users can always override this setting by toggling the "Use Agents" button in the UI when creating or editing an assistant. This environment variable only controls the initial default state. ## API Access PentAGI provides comprehensive programmatic access through both REST and GraphQL APIs, allowing you to integrate penetration testing workflows into your automation pipelines, CI/CD processes, and custom applications. ### Generating API Tokens API tokens are managed through the PentAGI web interface: 1. Navigate to **Settings** → **API Tokens** in the web UI 2. Click **Create Token** to generate a new API token 3. Configure token properties: - **Name** (optional): A descriptive name for the token - **Expiration Date**: When the token will expire (minimum 1 minute, maximum 3 years) 4. Click **Create** and **copy the token immediately** - it will only be shown once for security reasons 5. Use the token as a Bearer token in your API requests Each token is associated with your user account and inherits your role's permissions. ### Using API Tokens Include the API token in the `Authorization` header of your HTTP requests: ```bash # GraphQL API example curl -X POST https://your-pentagi-instance:8443/api/v1/graphql \ -H "Authorization: Bearer YOUR_API_TOKEN" \ -H "Content-Type: application/json" \ -d '{"query": "{ flows { id title status } }"}' # REST API example curl https://your-pentagi-instance:8443/api/v1/flows \ -H "Authorization: Bearer YOUR_API_TOKEN" ``` ### API Exploration and Testing PentAGI provides interactive documentation for exploring and testing API endpoints: #### GraphQL Playground Access the GraphQL Playground at `https://your-pentagi-instance:8443/api/v1/graphql/playground` 1. Click the **HTTP Headers** tab at the bottom 2. Add your authorization header: ```json { "Authorization": "Bearer YOUR_API_TOKEN" } ``` 3. Explore the schema, run queries, and test mutations interactively #### Swagger UI Access the REST API documentation at `https://your-pentagi-instance:8443/api/v1/swagger/index.html` 1. Click the **Authorize** button 2. Enter your token in the format: `Bearer YOUR_API_TOKEN` 3. Click **Authorize** to apply 4. Test endpoints directly from the Swagger UI ### Generating API Clients You can generate type-safe API clients for your preferred programming language using the schema files included with PentAGI: #### GraphQL Clients The GraphQL schema is available at: - **Web UI**: Navigate to Settings to download `schema.graphqls` - **Direct file**: `backend/pkg/graph/schema.graphqls` in the repository Generate clients using tools like: - **GraphQL Code Generator** (JavaScript/TypeScript): [https://the-guild.dev/graphql/codegen](https://the-guild.dev/graphql/codegen) - **genqlient** (Go): [https://github.com/Khan/genqlient](https://github.com/Khan/genqlient) - **Apollo iOS** (Swift): [https://www.apollographql.com/docs/ios](https://www.apollographql.com/docs/ios) #### REST API Clients The OpenAPI specification is available at: - **Swagger JSON**: `https://your-pentagi-instance:8443/api/v1/swagger/doc.json` - **Swagger YAML**: Available in `backend/pkg/server/docs/swagger.yaml` Generate clients using: - **OpenAPI Generator**: [https://openapi-generator.tech](https://openapi-generator.tech) ```bash openapi-generator-cli generate \ -i https://your-pentagi-instance:8443/api/v1/swagger/doc.json \ -g python \ -o ./pentagi-client ``` - **Swagger Codegen**: [https://github.com/swagger-api/swagger-codegen](https://github.com/swagger-api/swagger-codegen) ```bash swagger-codegen generate \ -i https://your-pentagi-instance:8443/api/v1/swagger/doc.json \ -l typescript-axios \ -o ./pentagi-client ``` - **swagger-typescript-api** (TypeScript): [https://github.com/acacode/swagger-typescript-api](https://github.com/acacode/swagger-typescript-api) ```bash npx swagger-typescript-api \ -p https://your-pentagi-instance:8443/api/v1/swagger/doc.json \ -o ./src/api \ -n pentagi-api.ts ``` ### API Usage Examples
Creating a New Flow (GraphQL) ```graphql mutation CreateFlow { createFlow( modelProvider: "openai" input: "Test the security of https://example.com" ) { id title status createdAt } } ```
Listing Flows (REST API) ```bash curl https://your-pentagi-instance:8443/api/v1/flows \ -H "Authorization: Bearer YOUR_API_TOKEN" \ | jq '.flows[] | {id, title, status}' ```
Python Client Example ```python import requests class PentAGIClient: def __init__(self, base_url, api_token): self.base_url = base_url self.headers = { "Authorization": f"Bearer {api_token}", "Content-Type": "application/json" } def create_flow(self, provider, target): query = """ mutation CreateFlow($provider: String!, $input: String!) { createFlow(modelProvider: $provider, input: $input) { id title status } } """ response = requests.post( f"{self.base_url}/api/v1/graphql", json={ "query": query, "variables": { "provider": provider, "input": target } }, headers=self.headers ) return response.json() def get_flows(self): response = requests.get( f"{self.base_url}/api/v1/flows", headers=self.headers ) return response.json() # Usage client = PentAGIClient( "https://your-pentagi-instance:8443", "your_api_token_here" ) # Create a new flow flow = client.create_flow("openai", "Scan https://example.com for vulnerabilities") print(f"Created flow: {flow}") # List all flows flows = client.get_flows() print(f"Total flows: {len(flows['flows'])}") ```
TypeScript Client Example ```typescript import axios, { AxiosInstance } from 'axios'; interface Flow { id: string; title: string; status: string; createdAt: string; } class PentAGIClient { private client: AxiosInstance; constructor(baseURL: string, apiToken: string) { this.client = axios.create({ baseURL: `${baseURL}/api/v1`, headers: { 'Authorization': `Bearer ${apiToken}`, 'Content-Type': 'application/json', }, }); } async createFlow(provider: string, input: string): Promise { const query = ` mutation CreateFlow($provider: String!, $input: String!) { createFlow(modelProvider: $provider, input: $input) { id title status createdAt } } `; const response = await this.client.post('/graphql', { query, variables: { provider, input }, }); return response.data.data.createFlow; } async getFlows(): Promise { const response = await this.client.get('/flows'); return response.data.flows; } async getFlow(flowId: string): Promise { const response = await this.client.get(`/flows/${flowId}`); return response.data; } } // Usage const client = new PentAGIClient( 'https://your-pentagi-instance:8443', 'your_api_token_here' ); // Create a new flow const flow = await client.createFlow( 'openai', 'Perform penetration test on https://example.com' ); console.log('Created flow:', flow); // List all flows const flows = await client.getFlows(); console.log(`Total flows: ${flows.length}`); ```
### Security Best Practices When working with API tokens: - **Never commit tokens to version control** - use environment variables or secrets management - **Rotate tokens regularly** - set appropriate expiration dates and create new tokens periodically - **Use separate tokens for different applications** - makes it easier to revoke access if needed - **Monitor token usage** - review API token activity in the Settings page - **Revoke unused tokens** - disable or delete tokens that are no longer needed - **Use HTTPS only** - never send API tokens over unencrypted connections ### Token Management - **View tokens**: See all your active tokens in Settings → API Tokens - **Edit tokens**: Update token names or revoke tokens - **Delete tokens**: Permanently remove tokens (this action cannot be undone) - **Token ID**: Each token has a unique ID that can be copied for reference The token list shows: - Token name (if provided) - Token ID (unique identifier) - Status (active/revoked/expired) - Creation date - Expiration date ### Custom LLM Provider Configuration When using custom LLM providers with the `LLM_SERVER_*` variables, you can fine-tune the reasoning format used in requests. > [!TIP] > For production-grade local deployments, consider using **vLLM** with **Qwen3.5-27B-FP8** for optimal performance. See our [comprehensive deployment guide](examples/guides/vllm-qwen35-27b-fp8.md) which includes hardware requirements, configuration templates ([thinking mode](examples/configs/vllm-qwen3.5-27b-fp8.provider.yml) and [non-thinking mode](examples/configs/vllm-qwen3.5-27b-fp8-no-think.provider.yml)), and performance benchmarks showing 13K TPS prompt processing on 4× RTX 5090 GPUs. | Variable | Default | Description | | ------------------------------- | ------- | --------------------------------------------------------------------------------------- | | `LLM_SERVER_URL` | | Base URL for the custom LLM API endpoint | | `LLM_SERVER_KEY` | | API key for the custom LLM provider | | `LLM_SERVER_MODEL` | | Default model to use (can be overridden in provider config) | | `LLM_SERVER_CONFIG_PATH` | | Path to the YAML configuration file for agent-specific models | | `LLM_SERVER_PROVIDER` | | Provider name prefix for model names (e.g., `openrouter`, `deepseek` for LiteLLM proxy) | | `LLM_SERVER_LEGACY_REASONING` | `false` | Controls reasoning format in API requests | | `LLM_SERVER_PRESERVE_REASONING` | `false` | Preserve reasoning content in multi-turn conversations (required by some providers) | The `LLM_SERVER_PROVIDER` setting is particularly useful when using **LiteLLM proxy**, which adds a provider prefix to model names. For example, when connecting to Moonshot API through LiteLLM, models like `kimi-2.5` become `moonshot/kimi-2.5`. By setting `LLM_SERVER_PROVIDER=moonshot`, you can use the same provider configuration file for both direct API access and LiteLLM proxy access without modifications. The `LLM_SERVER_LEGACY_REASONING` setting affects how reasoning parameters are sent to the LLM: - `false` (default): Uses modern format where reasoning is sent as a structured object with `max_tokens` parameter - `true`: Uses legacy format with string-based `reasoning_effort` parameter This setting is important when working with different LLM providers as they may expect different reasoning formats in their API requests. If you encounter reasoning-related errors with custom providers, try changing this setting. The `LLM_SERVER_PRESERVE_REASONING` setting controls whether reasoning content is preserved in multi-turn conversations: - `false` (default): Reasoning content is not preserved in conversation history - `true`: Reasoning content is preserved and sent in subsequent API calls This setting is required by some LLM providers (e.g., Moonshot) that return errors like "thinking is enabled but reasoning_content is missing in assistant tool call message" when reasoning content is not included in multi-turn conversations. Enable this setting if your provider requires reasoning content to be preserved. ### Ollama Provider Configuration PentAGI supports Ollama for both local LLM inference (zero-cost, enhanced privacy) and Ollama Cloud (managed service with free tier). #### Configuration Variables | Variable | Default | Description | | ----------------------------------- | ----------- | ----------------------------------------- | | `OLLAMA_SERVER_URL` | | URL of your Ollama server or Ollama Cloud | | `OLLAMA_SERVER_API_KEY` | | API key for Ollama Cloud authentication | | `OLLAMA_SERVER_MODEL` | | Default model for inference | | `OLLAMA_SERVER_CONFIG_PATH` | | Path to custom agent configuration file | | `OLLAMA_SERVER_PULL_MODELS_TIMEOUT` | `600` | Timeout for model downloads (seconds) | | `OLLAMA_SERVER_PULL_MODELS_ENABLED` | `false` | Auto-download models on startup | | `OLLAMA_SERVER_LOAD_MODELS_ENABLED` | `false` | Query server for available models | #### Ollama Cloud Configuration Ollama Cloud provides managed inference with a generous free tier and scalable paid plans. **Free Tier Setup (Single Model)** ```bash # Free tier allows one model at a time OLLAMA_SERVER_URL=https://ollama.com OLLAMA_SERVER_API_KEY=your_ollama_cloud_api_key OLLAMA_SERVER_MODEL=gpt-oss:120b # Example: OpenAI OSS 120B model ``` **Paid Tier Setup (Multi-Model with Pre-built Configuration)** For paid tiers supporting multiple concurrent models, use the pre-built Ollama Cloud configuration: ```bash # Using pre-built Ollama Cloud configuration (included in Docker image) OLLAMA_SERVER_URL=https://ollama.com OLLAMA_SERVER_API_KEY=your_ollama_cloud_api_key OLLAMA_SERVER_CONFIG_PATH=/opt/pentagi/conf/ollama-cloud.provider.yml ``` The pre-built `ollama-cloud.provider.yml` configuration includes optimized model assignments for all agent types: - **Simple/Assistant**: `nemotron-3-super:cloud` - Fast general-purpose model - **Primary Agent**: `qwen3-coder-next:cloud` - Advanced reasoning with high effort mode - **Coder/Pentester**: `qwen3-coder-next:cloud` - Specialized coding models - **Searcher**: `qwen3.5:397b-cloud` - Large context for information gathering - **Refiner/Refactor**: `glm-5:cloud` - High-quality text refinement - **Adviser/Enricher**: `minimax-m2.7:cloud` - Efficient advisory tasks - **Installer**: `devstral-2:123b-cloud` - Installation and setup tasks **Custom Configuration (Advanced)** To create your own agent configuration, mount a custom file from your host filesystem: ```bash # Using custom provider configuration OLLAMA_SERVER_URL=https://ollama.com OLLAMA_SERVER_API_KEY=your_ollama_cloud_api_key OLLAMA_SERVER_CONFIG_PATH=/opt/pentagi/conf/ollama.provider.yml # Mount custom configuration from host filesystem (in .env or docker-compose override) PENTAGI_OLLAMA_SERVER_CONFIG_PATH=/path/on/host/my-ollama-config.yml ``` The `PENTAGI_OLLAMA_SERVER_CONFIG_PATH` environment variable maps your host configuration file to `/opt/pentagi/conf/ollama.provider.yml` inside the container. **Example custom configuration** (`my-ollama-config.yml`): ```yaml primary_agent: model: "qwen3-coder-next:cloud" temperature: 1.0 top_p: 0.9 max_tokens: 32768 reasoning: effort: high coder: model: "qwen3-coder:32b" temperature: 1.0 max_tokens: 20480 ``` #### Local Ollama Configuration For self-hosted Ollama instances: ```bash # Basic local Ollama setup OLLAMA_SERVER_URL=http://localhost:11434 OLLAMA_SERVER_MODEL=llama3.1:8b-instruct-q8_0 # Production setup with auto-pull and model discovery OLLAMA_SERVER_URL=http://ollama-server:11434 OLLAMA_SERVER_PULL_MODELS_ENABLED=true OLLAMA_SERVER_PULL_MODELS_TIMEOUT=900 OLLAMA_SERVER_LOAD_MODELS_ENABLED=true # Using pre-built configurations from Docker image OLLAMA_SERVER_CONFIG_PATH=/opt/pentagi/conf/ollama-llama318b.provider.yml # or OLLAMA_SERVER_CONFIG_PATH=/opt/pentagi/conf/ollama-qwen332b-fp16-tc.provider.yml # or OLLAMA_SERVER_CONFIG_PATH=/opt/pentagi/conf/ollama-qwq32b-fp16-tc.provider.yml ``` **Performance Considerations:** - **Model Discovery** (`OLLAMA_SERVER_LOAD_MODELS_ENABLED=true`): Adds 1-2s startup latency querying Ollama API - **Auto-pull** (`OLLAMA_SERVER_PULL_MODELS_ENABLED=true`): First startup may take several minutes downloading models - **Pull timeout** (`OLLAMA_SERVER_PULL_MODELS_TIMEOUT=900`): 15 minutes in seconds - **Static Config**: Disable both flags and specify models in config file for fastest startup #### Creating Custom Ollama Models with Extended Context PentAGI requires models with larger context windows than the default Ollama configurations. You need to create custom models with increased `num_ctx` parameter through Modelfiles. While typical agent workflows consume around 64K tokens, PentAGI uses 110K context size for safety margin and handling complex penetration testing scenarios. **Important**: The `num_ctx` parameter can only be set during model creation via Modelfile - it cannot be changed after model creation or overridden at runtime. ##### Example: Qwen3 32B FP16 with Extended Context Create a Modelfile named `Modelfile_qwen3_32b_fp16_tc`: ```dockerfile FROM qwen3:32b-fp16 PARAMETER num_ctx 110000 PARAMETER temperature 0.3 PARAMETER top_p 0.8 PARAMETER min_p 0.0 PARAMETER top_k 20 PARAMETER repeat_penalty 1.1 ``` Build the custom model: ```bash ollama create qwen3:32b-fp16-tc -f Modelfile_qwen3_32b_fp16_tc ``` ##### Example: QwQ 32B FP16 with Extended Context Create a Modelfile named `Modelfile_qwq_32b_fp16_tc`: ```dockerfile FROM qwq:32b-fp16 PARAMETER num_ctx 110000 PARAMETER temperature 0.2 PARAMETER top_p 0.7 PARAMETER min_p 0.0 PARAMETER top_k 40 PARAMETER repeat_penalty 1.2 ``` Build the custom model: ```bash ollama create qwq:32b-fp16-tc -f Modelfile_qwq_32b_fp16_tc ``` > **Note**: The QwQ 32B FP16 model requires approximately **71.3 GB VRAM** for inference. Ensure your system has sufficient GPU memory before attempting to use this model. These custom models are referenced in the pre-built provider configuration files (`ollama-qwen332b-fp16-tc.provider.yml` and `ollama-qwq32b-fp16-tc.provider.yml`) that are included in the Docker image at `/opt/pentagi/conf/`. ### OpenAI Provider Configuration PentAGI integrates with OpenAI's comprehensive model lineup, featuring advanced reasoning capabilities with extended chain-of-thought, agentic models with enhanced tool integration, and specialized code models for security engineering. #### Configuration Variables | Variable | Default | Description | | -------------------- | --------------------------- | --------------------------- | | `OPEN_AI_KEY` | | API key for OpenAI services | | `OPEN_AI_SERVER_URL` | `https://api.openai.com/v1` | OpenAI API endpoint | #### Configuration Examples ```bash # Basic OpenAI setup OPEN_AI_KEY=your_openai_api_key OPEN_AI_SERVER_URL=https://api.openai.com/v1 # Using with proxy for enhanced security OPEN_AI_KEY=your_openai_api_key PROXY_URL=http://your-proxy:8080 ``` #### Supported Models PentAGI supports 31 OpenAI models with tool calling, streaming, reasoning modes, and prompt caching. Models marked with `*` are used in default configuration. **GPT-5.2 Series - Latest Flagship Agentic (December 2025)** | Model ID | Thinking | Price (Input/Output/Cache) | Use Case | | --------------------- | -------- | -------------------------- | ----------------------------------------------- | | `gpt-5.2`* | ✅ | $1.75/$14.00/$0.18 | Latest flagship with enhanced reasoning and tool integration, autonomous security research | | `gpt-5.2-pro` | ✅ | $21.00/$168.00/$0.00 | Premium version with superior agentic coding, mission-critical security research, zero-day discovery | | `gpt-5.2-codex` | ✅ | $1.75/$14.00/$0.18 | Most advanced code-specialized, context compaction, strong cybersecurity capabilities | **GPT-5/5.1 Series - Advanced Agentic Models** | Model ID | Thinking | Price (Input/Output/Cache) | Use Case | | --------------------- | -------- | -------------------------- | ----------------------------------------------- | | `gpt-5` | ✅ | $1.25/$10.00/$0.13 | Premier agentic with advanced reasoning, autonomous security research, exploit chain development | | `gpt-5.1` | ✅ | $1.25/$10.00/$0.13 | Enhanced agentic with adaptive reasoning, balanced penetration testing with strong tool coordination | | `gpt-5-pro` | ✅ | $15.00/$120.00/$0.00 | Premium version with major reasoning improvements, reduced hallucinations, critical security operations | | `gpt-5-mini` | ✅ | $0.25/$2.00/$0.03 | Efficient balancing speed and intelligence, automated vulnerability analysis, exploit generation | | `gpt-5-nano` | ✅ | $0.05/$0.40/$0.01 | Fastest for high-throughput scanning, reconnaissance, bulk vulnerability detection | **GPT-5/5.1 Codex Series - Code-Specialized** | Model ID | Thinking | Price (Input/Output/Cache) | Use Case | | --------------------- | -------- | -------------------------- | ----------------------------------------------- | | `gpt-5.1-codex-max` | ✅ | $1.25/$10.00/$0.13 | Enhanced reasoning for sophisticated coding, proven CVE findings, systematic exploit development | | `gpt-5.1-codex` | ✅ | $1.25/$10.00/$0.13 | Standard code-optimized with strong reasoning, exploit generation, vulnerability analysis | | `gpt-5-codex` | ✅ | $1.25/$10.00/$0.13 | Foundational code-specialized, vulnerability scanning, basic exploit generation | | `gpt-5.1-codex-mini` | ✅ | $0.25/$2.00/$0.03 | Compact high-performance, 4x higher capacity, rapid vulnerability detection | | `codex-mini-latest` | ✅ | $1.50/$6.00/$0.38 | Latest compact code model, automated code review, basic vulnerability analysis | **GPT-4.1 Series - Enhanced Intelligence** | Model ID | Thinking | Price (Input/Output/Cache) | Use Case | | --------------------- | -------- | -------------------------- | ----------------------------------------------- | | `gpt-4.1` | ❌ | $2.00/$8.00/$0.50 | Enhanced flagship with superior function calling, complex threat analysis, sophisticated exploit development | | `gpt-4.1-mini`* | ❌ | $0.40/$1.60/$0.10 | Balanced performance with improved efficiency, routine security assessments, automated code analysis | | `gpt-4.1-nano` | ❌ | $0.10/$0.40/$0.03 | Ultra-fast lightweight, bulk security scanning, rapid reconnaissance, continuous monitoring | **GPT-4o Series - Multimodal Flagship** | Model ID | Thinking | Price (Input/Output/Cache) | Use Case | | --------------------- | -------- | -------------------------- | ----------------------------------------------- | | `gpt-4o` | ❌ | $2.50/$10.00/$1.25 | Multimodal flagship with vision, image analysis, web UI assessment, multi-tool orchestration | | `gpt-4o-mini` | ❌ | $0.15/$0.60/$0.08 | Compact multimodal with strong function calling, high-frequency scanning, cost-effective bulk operations | **o-Series - Advanced Reasoning Models** | Model ID | Thinking | Price (Input/Output/Cache) | Use Case | | --------------------- | -------- | -------------------------- | ----------------------------------------------- | | `o4-mini`* | ✅ | $1.10/$4.40/$0.28 | Next-gen reasoning with enhanced speed, methodical security assessments, systematic exploit development | | `o3`* | ✅ | $2.00/$8.00/$0.50 | Advanced reasoning powerhouse, multi-stage attack chains, deep vulnerability analysis | | `o3-mini` | ✅ | $1.10/$4.40/$0.55 | Compact reasoning with extended thinking, step-by-step attack planning, logical vulnerability chaining | | `o1` | ✅ | $15.00/$60.00/$7.50 | Premier reasoning with maximum depth, advanced penetration testing, novel exploit research | | `o3-pro` | ✅ | $20.00/$80.00/$0.00 | Most advanced reasoning, 80% cheaper than o1-pro, zero-day research, critical security investigations | | `o1-pro` | ✅ | $150.00/$600.00/$0.00 | Previous-gen premium reasoning, exhaustive security analysis, mission-critical challenges | **Prices**: Per 1M tokens. Reasoning models include thinking tokens in output pricing. > [!WARNING] > **GPT-5* Models - Trusted Access Required** > > All GPT-5 series models (`gpt-5`, `gpt-5.1`, `gpt-5.2`, `gpt-5-pro`, `gpt-5.2-pro`, and all Codex variants) work **unstably with PentAGI** and may trigger OpenAI's cybersecurity safety mechanisms without verified access. > > **To use GPT-5* models reliably:** > 1. **Individual users**: Verify your identity at [chatgpt.com/cyber](https://chatgpt.com/cyber) > 2. **Enterprise teams**: Request trusted access through your OpenAI representative > 3. **Security researchers**: Apply for the [Cybersecurity Grant Program](https://openai.com/form/cybersecurity-grant-program/) (includes $10M in API credits) > > **Recommended alternatives without verification:** > - Use `o-series` models (o3, o4-mini, o1) for reasoning tasks > - Use `gpt-4.1` series for general intelligence and function calling > - All o-series and gpt-4.x models work reliably without special access **Reasoning Effort Levels**: - **High**: Maximum reasoning depth (refiner - o3 with high effort) - **Medium**: Balanced reasoning (primary_agent, assistant, reflector - o4-mini/o3 with medium effort) - **Low**: Efficient targeted reasoning (coder, installer, pentester - o3/o4-mini with low effort; adviser - gpt-5.2 with low effort) **Key Features**: - **Extended Reasoning**: o-series models with chain-of-thought for complex security analysis - **Agentic Intelligence**: GPT-5/5.1/5.2 series with enhanced tool integration and autonomous capabilities - **Prompt Caching**: Cost reduction on repeated context (10-50% of input price) - **Code Specialization**: Dedicated Codex models for vulnerability discovery and exploit development - **Multimodal Support**: GPT-4o series for vision-based security assessments - **Tool Calling**: Robust function calling across all models for pentesting tool orchestration - **Streaming**: Real-time response streaming for interactive workflows - **Proven Track Record**: Industry-leading models with CVE discoveries and real-world security applications ### Anthropic Provider Configuration PentAGI integrates with Anthropic's Claude models, featuring advanced extended thinking capabilities, exceptional safety mechanisms, and sophisticated understanding of complex security contexts with prompt caching. #### Configuration Variables | Variable | Default | Description | | ---------------------- | ------------------------------ | ------------------------------ | | `ANTHROPIC_API_KEY` | | API key for Anthropic services | | `ANTHROPIC_SERVER_URL` | `https://api.anthropic.com/v1` | Anthropic API endpoint | #### Configuration Examples ```bash # Basic Anthropic setup ANTHROPIC_API_KEY=your_anthropic_api_key ANTHROPIC_SERVER_URL=https://api.anthropic.com/v1 # Using with proxy for secure environments ANTHROPIC_API_KEY=your_anthropic_api_key PROXY_URL=http://your-proxy:8080 ``` #### Supported Models PentAGI supports 10 Claude models with tool calling, streaming, extended thinking, adaptive thinking, and prompt caching. Models marked with `*` are used in default configuration. **Claude 4 Series - Latest Models (2025-2026)** | Model ID | Thinking | Release Date | Price (Input/Output/Cache R/W) | Use Case | | ------------------------ | -------- | ------------ | ------------------------------ | ----------------------------------------------- | | `claude-opus-4-6`* | ✅ | May 2025 | $5.00/$25.00/$0.50/$6.25 | Most intelligent model for autonomous agents and coding. Extended + adaptive thinking for complex exploit development, multi-stage attack simulation | | `claude-sonnet-4-6`* | ✅ | Aug 2025 | $3.00/$15.00/$0.30/$3.75 | Best speed/intelligence balance with adaptive thinking. Multi-phase security assessments, intelligent vulnerability analysis, real-time threat hunting | | `claude-haiku-4-5`* | ✅ | Oct 2025 | $1.00/$5.00/$0.10/$1.25 | Fastest model with near-frontier intelligence. High-frequency scanning, real-time monitoring, bulk automated testing | **Legacy Models - Still Supported** | Model ID | Thinking | Release Date | Price (Input/Output/Cache R/W) | Use Case | | ------------------------ | -------- | ------------ | ------------------------------ | ----------------------------------------------- | | `claude-sonnet-4-5` | ✅ | Sep 2025 | $3.00/$15.00/$0.30/$3.75 | State-of-the-art reasoning (superseded by 4-6). Sophisticated penetration testing, advanced threat analysis | | `claude-opus-4-5` | ✅ | Nov 2025 | $5.00/$25.00/$0.50/$6.25 | Ultimate reasoning (superseded by opus-4-6). Critical security research, zero-day discovery, red team operations | | `claude-opus-4-1` | ✅ | Aug 2025 | $15.00/$75.00/$1.50/$18.75 | Advanced reasoning (superseded). Complex penetration testing, sophisticated threat modeling | | `claude-sonnet-4-0` | ✅ | May 2025 | $3.00/$15.00/$0.30/$3.75 | High-performance reasoning (superseded). Complex threat modeling, multi-tool coordination | | `claude-opus-4-0` | ✅ | May 2025 | $15.00/$75.00/$1.50/$18.75 | First generation Opus (superseded). Multi-step exploit development, autonomous pentesting workflows | **Deprecated Models - Migrate to Current Models** | Model ID | Thinking | Release Date | Price (Input/Output/Cache R/W) | Notes | | ---------------------------- | -------- | ------------ | ------------------------------ | -------------------------------------------- | | `claude-3-haiku-20240307` | ❌ | Mar 2024 | $0.25/$1.25/$0.03/$0.30 | Will be retired April 19, 2026. Migrate to claude-haiku-4-5 | **Prices**: Per 1M tokens. Cache pricing includes both Read and Write costs. **Extended Thinking Configuration**: - **Max Tokens 4096**: Generator (claude-opus-4-6) for maximum reasoning depth on complex exploit development - **Max Tokens 2048**: Coder (claude-sonnet-4-6) for balanced code analysis and vulnerability research - **Max Tokens 1024**: Primary agent, assistant, refiner, adviser, reflector, searcher, installer, pentester for focused reasoning on specific tasks - **Extended Thinking**: All Claude 4.5+ and 4.6 models support configurable extended thinking for deep reasoning tasks **Key Features**: - **Extended Thinking**: All Claude 4.5+ and 4.6 models with configurable chain-of-thought reasoning depths for complex security analysis - **Adaptive Thinking**: Claude 4.6 series (Opus/Sonnet) dynamically adjusts reasoning depth based on task complexity for optimal performance - **Prompt Caching**: Significant cost reduction with separate read/write pricing (10% read, 125% write of input) - **Extended Context Window**: 200K tokens standard, up to 1M tokens (beta) for Claude Opus/Sonnet 4.6 for comprehensive codebase analysis - **Tool Calling**: Robust function calling with exceptional accuracy for security tool orchestration - **Streaming**: Real-time response streaming for interactive penetration testing workflows - **Safety-First Design**: Built-in safety mechanisms ensuring responsible security testing practices - **Multimodal Support**: Vision capabilities in latest models for screenshot analysis and UI security assessment - **Constitutional AI**: Advanced safety training providing reliable and ethical security guidance ### Google AI (Gemini) Provider Configuration PentAGI integrates with Google's Gemini models through the Google AI API, offering state-of-the-art multimodal reasoning capabilities with extended thinking and context caching. #### Configuration Variables | Variable | Default | Description | | ------------------- | ------------------------------------------- | ------------------------------ | | `GEMINI_API_KEY` | | API key for Google AI services | | `GEMINI_SERVER_URL` | `https://generativelanguage.googleapis.com` | Google AI API endpoint | #### Configuration Examples ```bash # Basic Gemini setup GEMINI_API_KEY=your_gemini_api_key GEMINI_SERVER_URL=https://generativelanguage.googleapis.com # Using with proxy GEMINI_API_KEY=your_gemini_api_key PROXY_URL=http://your-proxy:8080 ``` #### Supported Models PentAGI supports 13 Gemini models with tool calling, streaming, thinking modes, and context caching. Models marked with `*` are used in default configuration. **Gemini 3.1 Series - Latest Flagship (February 2026)** | Model ID | Thinking | Context | Price (Input/Output/Cache) | Use Case | | ------------------------------------- | -------- | ------- | -------------------------- | ----------------------------------------------- | | `gemini-3.1-pro-preview`* | ✅ | 1M | $2.00/$12.00/$0.20 | Latest flagship with refined thinking, improved token efficiency, optimized for software engineering and agentic workflows | | `gemini-3.1-pro-preview-customtools` | ✅ | 1M | $2.00/$12.00/$0.20 | Custom tools endpoint optimized for bash and custom tools (view_file, search_code) prioritization | | `gemini-3.1-flash-lite-preview`* | ✅ | 1M | $0.25/$1.50/$0.03 | Most cost-efficient with fastest performance for high-volume agentic tasks and low-latency applications | **Gemini 3 Series (⚠️ gemini-3-pro-preview DEPRECATED - Shutdown March 9, 2026)** | Model ID | Thinking | Context | Price (Input/Output/Cache) | Use Case | | ------------------------------------- | -------- | ------- | -------------------------- | ----------------------------------------------- | | `gemini-3-pro-preview` | ✅ | 1M | $2.00/$12.00/$0.20 | ⚠️ DEPRECATED - Migrate to gemini-3.1-pro-preview before March 9, 2026 | | `gemini-3-flash-preview`* | ✅ | 1M | $0.50/$3.00/$0.05 | Frontier intelligence with superior search grounding, high-throughput security scanning | **Gemini 2.5 Series - Advanced Thinking Models** | Model ID | Thinking | Context | Price (Input/Output/Cache) | Use Case | | ---------------------------------------- | -------- | ------- | -------------------------- | ----------------------------------------------- | | `gemini-2.5-pro` | ✅ | 1M | $1.25/$10.00/$0.13 | State-of-the-art for complex coding and reasoning, sophisticated threat modeling | | `gemini-2.5-flash` | ✅ | 1M | $0.30/$2.50/$0.03 | First hybrid reasoning model with thinking budgets, best price-performance for large-scale assessments | | `gemini-2.5-flash-lite` | ✅ | 1M | $0.10/$0.40/$0.01 | Smallest and most cost-effective for at-scale usage, high-throughput scanning | | `gemini-2.5-flash-lite-preview-09-2025` | ✅ | 1M | $0.10/$0.40/$0.01 | Latest preview optimized for cost-efficiency, high throughput, and quality | **Gemini 2.0 Series - Balanced Multimodal for Agents** | Model ID | Thinking | Context | Price (Input/Output/Cache) | Use Case | | ------------------------------------- | -------- | ------- | -------------------------- | ----------------------------------------------- | | `gemini-2.0-flash` | ❌ | 1M | $0.10/$0.40/$0.03 | Balanced multimodal built for agents era, diverse security tasks and real-time monitoring | | `gemini-2.0-flash-lite` | ❌ | 1M | $0.08/$0.30/$0.00 | Lightweight for continuous monitoring, basic scanning, automated alert processing | **Specialized Open-Source Models (Free)** | Model ID | Thinking | Context | Price (Input/Output/Cache) | Use Case | | ------------------------------------- | -------- | ------- | -------------------------- | ----------------------------------------------- | | `gemma-3-27b-it` | ❌ | 128K | Free/Free/Free | Open-source from Gemini tech, on-premises security operations, privacy-sensitive testing | | `gemma-3n-4b-it` | ❌ | 128K | Free/Free/Free | Efficient for edge devices (mobile/laptops/tablets), offline vulnerability scanning | **Prices**: Per 1M tokens (Standard Paid tier). Context window is input token limit. > [!WARNING] > **Gemini 3 Pro Preview Deprecation** > > `gemini-3-pro-preview` will be **shut down on March 9, 2026**. Migrate to `gemini-3.1-pro-preview` to avoid service disruption. The new model offers: > > - Refined performance and reliability > - Improved thinking and token efficiency > - Better grounded, factually consistent responses > - Enhanced software engineering behavior **Key Features**: - **Extended Thinking**: Step-by-step reasoning for complex security analysis (all Gemini 3.x and 2.5 series) - **Context Caching**: Significant cost reduction on repeated context (10-90% of input price) - **Ultra-Long Context**: 1M tokens for comprehensive codebase analysis and documentation review - **Multimodal Support**: Text, image, video, audio, and PDF processing for comprehensive assessments - **Tool Calling**: Seamless integration with 20+ pentesting tools via function calling - **Streaming**: Real-time response streaming for interactive security workflows - **Code Execution**: Built-in code execution for offensive tool testing and exploit validation - **Search Grounding**: Google Search integration for threat intelligence and CVE research - **File Search**: Document retrieval and RAG capabilities for knowledge-based assessments - **Batch API**: 50% cost reduction for non-real-time batch processing **Reasoning Effort Levels**: - **High**: Maximum thinking depth for complex multi-step analysis (generator) - **Medium**: Balanced reasoning for general agentic tasks (primary_agent, assistant, refiner, adviser) - **Low**: Efficient thinking for focused tasks (coder, installer, pentester) ### AWS Bedrock Provider Configuration PentAGI integrates with Amazon Bedrock, offering access to 20+ foundation models from leading AI companies including Anthropic, Amazon, Cohere, DeepSeek, OpenAI, Qwen, Mistral, and Moonshot. #### Configuration Variables | Variable | Default | Description | | --------------------------- | ----------- | --------------------------------------------------------------------------------------------------- | | `BEDROCK_REGION` | `us-east-1` | AWS region for Bedrock service | | `BEDROCK_DEFAULT_AUTH` | `false` | Use AWS SDK default credential chain (environment, EC2 role, ~/.aws/credentials) - highest priority | | `BEDROCK_BEARER_TOKEN` | | Bearer token authentication - priority over static credentials | | `BEDROCK_ACCESS_KEY_ID` | | AWS access key ID for static credentials | | `BEDROCK_SECRET_ACCESS_KEY` | | AWS secret access key for static credentials | | `BEDROCK_SESSION_TOKEN` | | AWS session token for temporary credentials (optional, used with static credentials) | | `BEDROCK_SERVER_URL` | | Custom Bedrock endpoint (VPC endpoints, local testing) | **Authentication Priority**: `BEDROCK_DEFAULT_AUTH` → `BEDROCK_BEARER_TOKEN` → `BEDROCK_ACCESS_KEY_ID`+`BEDROCK_SECRET_ACCESS_KEY` #### Configuration Examples ```bash # Recommended: Default AWS SDK authentication (EC2/ECS/Lambda roles) BEDROCK_REGION=us-east-1 BEDROCK_DEFAULT_AUTH=true # Bearer token authentication (AWS STS, custom auth) BEDROCK_REGION=us-east-1 BEDROCK_BEARER_TOKEN=your_bearer_token # Static credentials (development, testing) BEDROCK_REGION=us-east-1 BEDROCK_ACCESS_KEY_ID=your_aws_access_key BEDROCK_SECRET_ACCESS_KEY=your_aws_secret_key # With proxy and custom endpoint BEDROCK_REGION=us-east-1 BEDROCK_DEFAULT_AUTH=true BEDROCK_SERVER_URL=https://bedrock-runtime.us-east-1.vpce-xxx.amazonaws.com PROXY_URL=http://your-proxy:8080 ``` #### Supported Models PentAGI supports 21 AWS Bedrock models with tool calling, streaming, and multimodal capabilities. Models marked with `*` are used in default configuration. | Model ID | Provider | Thinking | Multimodal | Price (Input/Output) | Use Case | | ------------------------------------------------ | --------------- | -------- | ---------- | -------------------- | --------------------------------------- | | `us.amazon.nova-2-lite-v1:0` | Amazon Nova | ❌ | ✅ | $0.33/$2.75 | Adaptive reasoning, efficient thinking | | `us.amazon.nova-premier-v1:0` | Amazon Nova | ❌ | ✅ | $2.50/$12.50 | Complex reasoning, advanced analysis | | `us.amazon.nova-pro-v1:0` | Amazon Nova | ❌ | ✅ | $0.80/$3.20 | Balanced accuracy, speed, cost | | `us.amazon.nova-lite-v1:0` | Amazon Nova | ❌ | ✅ | $0.06/$0.24 | Fast processing, high-volume operations | | `us.amazon.nova-micro-v1:0` | Amazon Nova | ❌ | ❌ | $0.035/$0.14 | Ultra-low latency, real-time monitoring | | `us.anthropic.claude-opus-4-6-v1`* | Anthropic | ✅ | ✅ | $5.00/$25.00 | World-class coding, enterprise agents | | `us.anthropic.claude-sonnet-4-6` | Anthropic | ✅ | ✅ | $3.00/$15.00 | Frontier intelligence, enterprise scale | | `us.anthropic.claude-opus-4-5-20251101-v1:0` | Anthropic | ✅ | ✅ | $5.00/$25.00 | Multi-day software development | | `us.anthropic.claude-haiku-4-5-20251001-v1:0`* | Anthropic | ✅ | ✅ | $1.00/$5.00 | Near-frontier performance, high speed | | `us.anthropic.claude-sonnet-4-5-20250929-v1:0`* | Anthropic | ✅ | ✅ | $3.00/$15.00 | Real-world agents, coding excellence | | `us.anthropic.claude-sonnet-4-20250514-v1:0` | Anthropic | ✅ | ✅ | $3.00/$15.00 | Balanced performance, production-ready | | `us.anthropic.claude-3-5-haiku-20241022-v1:0` | Anthropic | ❌ | ❌ | $0.80/$4.00 | Fastest model, cost-effective scanning | | `cohere.command-r-plus-v1:0` | Cohere | ❌ | ❌ | $3.00/$15.00 | Large-scale operations, superior RAG | | `deepseek.v3.2` | DeepSeek | ❌ | ❌ | $0.58/$1.68 | Long-context reasoning, efficiency | | `openai.gpt-oss-120b-1:0`* | OpenAI (OSS) | ✅ | ❌ | $0.15/$0.60 | Strong reasoning, scientific analysis | | `openai.gpt-oss-20b-1:0` | OpenAI (OSS) | ✅ | ❌ | $0.07/$0.30 | Efficient coding, software development | | `qwen.qwen3-next-80b-a3b` | Qwen | ❌ | ❌ | $0.15/$1.20 | Ultra-long context, flagship reasoning | | `qwen.qwen3-32b-v1:0` | Qwen | ❌ | ❌ | $0.15/$0.60 | Balanced reasoning, research use cases | | `qwen.qwen3-coder-30b-a3b-v1:0` | Qwen | ❌ | ❌ | $0.15/$0.60 | Vibe coding, natural-language first | | `qwen.qwen3-coder-next` | Qwen | ❌ | ❌ | $0.45/$1.80 | Tool use, function calling optimized | | `mistral.mistral-large-3-675b-instruct` | Mistral | ❌ | ✅ | $4.00/$12.00 | Advanced multimodal, long-context | | `moonshotai.kimi-k2.5` | Moonshot | ❌ | ✅ | $0.60/$3.00 | Vision, language, code in one model | **Prices**: Per 1M tokens. Models with thinking/reasoning support additional compute costs during reasoning phase. #### Tested but Incompatible Models Some AWS Bedrock models were tested but are **not supported** due to technical limitations: | Model Family | Reason for Incompatibility | | ------------------------- | ----------------------------------------------------------------------------------------- | | **GLM (Z.AI)** | Tool calling format incompatible with Converse API (expects string instead of JSON) | | **AI21 Jamba** | Severe rate limits (1-2 req/min) prevent reliable testing and production use | | **Meta Llama 3.3/3.1** | Unstable tool call result processing, causes unexpected failures in multi-turn workflows | | **Mistral Magistral** | Tool calling not supported by the model | | **Moonshot K2-Thinking** | Unstable streaming behavior with tool calls, unreliable in production | | **Qwen3-VL** | Unstable streaming with tool calling, multimodal + tools combination fails intermittently | > [!IMPORTANT] > **Rate Limits & Quota Management** > > Default AWS Bedrock quotas for Claude models are **extremely restrictive** (2-20 requests/minute for new accounts). For production penetration testing: > > 1. **Request quota increases** through AWS Service Quotas console for models you plan to use > 2. **Use Amazon Nova models** - higher default quotas and excellent performance > 3. **Enable provisioned throughput** for consistent high-volume testing > 4. **Monitor usage** - AWS throttles aggressively at quota limits > > Without quota increases, expect frequent delays and workflow interruptions. > [!WARNING] > **Converse API Requirements** > > PentAGI uses Amazon Bedrock **Converse API** for unified model access. All supported models require: > > - ✅ Converse/ConverseStream API support > - ✅ Tool use (function calling) for penetration testing workflows > - ✅ Streaming tool use for real-time feedback > > Verify model capabilities at: [AWS Bedrock Model Features](https://docs.aws.amazon.com/bedrock/latest/userguide/conversation-inference-supported-models-features.html) **Key Features**: - **Automatic Prompt Caching**: 40-70% cost reduction on repeated context (Claude 4.x models) - **Extended Thinking**: Step-by-step reasoning for complex security analysis (Claude, DeepSeek R1, OpenAI GPT) - **Multimodal Analysis**: Process screenshots, diagrams, video for comprehensive testing (Nova, Claude, Mistral, Kimi) - **Tool Calling**: Seamless integration with 20+ pentesting tools via function calling - **Streaming**: Real-time response streaming for interactive security assessment workflows ### DeepSeek Provider Configuration PentAGI integrates with DeepSeek, providing access to advanced AI models with strong reasoning, coding capabilities, and context caching at competitive prices. #### Configuration Variables | Variable | Default Value | Description | | --------------------- | -------------------------- | --------------------------------------------------- | | `DEEPSEEK_API_KEY` | | DeepSeek API key for authentication | | `DEEPSEEK_SERVER_URL` | `https://api.deepseek.com` | DeepSeek API endpoint URL | | `DEEPSEEK_PROVIDER` | | Provider prefix for LiteLLM integration (optional) | #### Configuration Examples ```bash # Direct API usage DEEPSEEK_API_KEY=your_deepseek_api_key DEEPSEEK_SERVER_URL=https://api.deepseek.com # With LiteLLM proxy DEEPSEEK_API_KEY=your_litellm_key DEEPSEEK_SERVER_URL=http://litellm-proxy:4000 DEEPSEEK_PROVIDER=deepseek # Adds prefix to model names (deepseek/deepseek-chat) for LiteLLM ``` #### Supported Models PentAGI supports 2 DeepSeek-V3.2 models with tool calling, streaming, thinking modes, and context caching. Both models are used in default configuration. | Model ID | Thinking | Context | Max Output | Price (Input/Output/Cache) | Use Case | | --------------------- | -------- | ------- | ---------- | -------------------------- | ----------------------------------------------- | | `deepseek-chat`* | ❌ | 128K | 8K | $0.28/$0.42/$0.03 | General dialogue, code generation, tool calling | | `deepseek-reasoner`* | ✅ | 128K | 64K | $0.28/$0.42/$0.03 | Advanced reasoning, complex logic, security analysis | **Prices**: Per 1M tokens. Cache pricing is for prompt caching (10% of input cost). Models with thinking support include reinforcement learning chain-of-thought reasoning. **Key Features**: - **Automatic Prompt Caching**: 40-60% cost reduction on repeated context (10% of input price) - **Extended Thinking**: Reinforcement learning CoT for complex security analysis (deepseek-reasoner) - **Strong Coding**: Optimized for code generation and exploit development - **Tool Calling**: Seamless integration with 20+ pentesting tools via function calling - **Streaming**: Real-time response streaming for interactive workflows - **Multilingual**: Strong Chinese and English support - **Additional Features**: JSON Output, Chat Prefix Completion, FIM (Fill-in-the-Middle) Completion **LiteLLM Integration**: Set `DEEPSEEK_PROVIDER=deepseek` to enable model name prefixing when using default PentAGI configurations with LiteLLM proxy. Leave empty for direct API usage. ### GLM Provider Configuration PentAGI integrates with GLM from Zhipu AI (Z.AI), providing advanced language models with MoE architecture, strong reasoning, and agentic capabilities developed by Tsinghua University. #### Configuration Variables | Variable | Default Value | Description | | ----------------- | ------------------------------- | ---------------------------------------------------------- | | `GLM_API_KEY` | | GLM API key for authentication | | `GLM_SERVER_URL` | `https://api.z.ai/api/paas/v4` | GLM API endpoint URL (international) | | `GLM_PROVIDER` | | Provider prefix for LiteLLM integration (optional) | #### Configuration Examples ```bash # Direct API usage (international endpoint) GLM_API_KEY=your_glm_api_key GLM_SERVER_URL=https://api.z.ai/api/paas/v4 # Alternative endpoints GLM_SERVER_URL=https://open.bigmodel.cn/api/paas/v4 # China GLM_SERVER_URL=https://api.z.ai/api/coding/paas/v4 # Coding-specific # With LiteLLM proxy GLM_API_KEY=your_litellm_key GLM_SERVER_URL=http://litellm-proxy:4000 GLM_PROVIDER=zai # Adds prefix to model names (zai/glm-4) for LiteLLM ``` #### Supported Models PentAGI supports 12 GLM models with tool calling, streaming, thinking modes, and prompt caching. Models marked with `*` are used in default configuration. **GLM-5 Series - Flagship MoE (744B/40B active)** | Model ID | Thinking | Context | Max Output | Price (Input/Output/Cache) | Use Case | | ----------------------- | ------------- | ------- | ---------- | -------------------------- | ----------------------------------------------- | | `glm-5`* | ✅ Forced | 200K | 128K | $1.00/$3.20/$0.20 | Flagship agentic engineering, complex multi-stage tasks | | `glm-5-code`† | ✅ Forced | 200K | 128K | $1.20/$5.00/$0.30 | Code-specialized, exploit development (requires Coding Plan) | **GLM-4.7 Series - Premium with Interleaved Thinking** | Model ID | Thinking | Context | Max Output | Price (Input/Output/Cache) | Use Case | | ----------------------- | ------------- | ------- | ---------- | -------------------------- | ----------------------------------------------- | | `glm-4.7`* | ✅ Forced | 200K | 128K | $0.60/$2.20/$0.11 | Premium with thinking before each response/tool call | | `glm-4.7-flashx`* | ✅ Hybrid | 200K | 128K | $0.07/$0.40/$0.01 | High-speed with priority GPU, best price/performance | | `glm-4.7-flash` | ✅ Hybrid | 200K | 128K | Free/Free/Free | Free ~30B SOTA model, 1 concurrent request | **GLM-4.6 Series - Balanced with Auto-Thinking** | Model ID | Thinking | Context | Max Output | Price (Input/Output/Cache) | Use Case | | ----------------------- | ------------- | ------- | ---------- | -------------------------- | ----------------------------------------------- | | `glm-4.6` | ✅ Auto | 200K | 128K | $0.60/$2.20/$0.11 | Balanced, streaming tool calls, 30% token efficient | **GLM-4.5 Series - Unified Reasoning/Coding/Agents** | Model ID | Thinking | Context | Max Output | Price (Input/Output/Cache) | Use Case | | ----------------------- | ------------- | ------- | ---------- | -------------------------- | ----------------------------------------------- | | `glm-4.5` | ✅ Auto | 128K | 96K | $0.60/$2.20/$0.11 | Unified model, MoE 355B/32B active | | `glm-4.5-x` | ✅ Auto | 128K | 96K | $2.20/$8.90/$0.45 | Ultra-fast premium, lowest latency | | `glm-4.5-air`* | ✅ Auto | 128K | 96K | $0.20/$1.10/$0.03 | Cost-effective, MoE 106B/12B, best price/quality | | `glm-4.5-airx` | ✅ Auto | 128K | 96K | $1.10/$4.50/$0.22 | Accelerated Air with priority GPU | | `glm-4.5-flash` | ✅ Auto | 128K | 96K | Free/Free/Free | Free with reasoning/coding/agents support | **GLM-4 Legacy - Dense Architecture** | Model ID | Thinking | Context | Max Output | Price (Input/Output/Cache) | Use Case | | ----------------------- | ------------- | ------- | ---------- | -------------------------- | ----------------------------------------------- | | `glm-4-32b-0414-128k` | ❌ | 128K | 16K | $0.10/$0.10/$0.00 | Ultra-budget dense 32B, high-volume parsing | **Prices**: Per 1M tokens. Cache pricing is for prompt caching. † Model requires **Coding Plan subscription**. > [!WARNING] > **Coding Plan Requirement** > > The `glm-5-code` model requires an active **Coding Plan subscription**. Attempting to use this model without the subscription will result in: > > ``` > API returned unexpected status code: 403: You do not have permission to access glm-5-code > ``` > > For code-specialized tasks without Coding Plan, use `glm-5` (general flagship) or `glm-4.7` (premium with interleaved thinking) instead. **Thinking Modes**: - **Forced**: Model always uses thinking mode before responding (GLM-5, GLM-4.7) - **Hybrid**: Model intelligently decides when to use thinking (GLM-4.7-FlashX, GLM-4.7-Flash) - **Auto**: Model automatically determines when reasoning is needed (GLM-4.6, GLM-4.5 series) **Key Features**: - **Prompt Caching**: Significant cost reduction on repeated context (cached input pricing shown) - **Interleaved Thinking**: GLM-4.7 thinks before each response and tool call with preserved reasoning across multi-turn dialogues - **Ultra-Long Context**: 200K tokens for GLM-5 and GLM-4.7/4.6 series for massive codebase analysis - **MoE Architecture**: Efficient 744B parameters with 40B active (GLM-5), 355B/32B (GLM-4.5), 106B/12B (GLM-4.5-Air) - **Tool Calling**: Seamless integration with 20+ pentesting tools via function calling - **Streaming**: Real-time response streaming with streaming tool calls support (GLM-4.6+) - **Multilingual**: Exceptional Chinese and English NLP capabilities - **Free Options**: GLM-4.7-Flash and GLM-4.5-Flash for prototyping and experimentation **LiteLLM Integration**: Set `GLM_PROVIDER=zai` to enable model name prefixing when using default PentAGI configurations with LiteLLM proxy. Leave empty for direct API usage. ### Kimi Provider Configuration PentAGI integrates with Kimi from Moonshot AI, providing ultra-long context models with multimodal capabilities perfect for analyzing extensive codebases and documentation. #### Configuration Variables | Variable | Default Value | Description | | ------------------ | -----------------------------| --------------------------------------------------- | | `KIMI_API_KEY` | | Kimi API key for authentication | | `KIMI_SERVER_URL` | `https://api.moonshot.ai/v1` | Kimi API endpoint URL (international) | | `KIMI_PROVIDER` | | Provider prefix for LiteLLM integration (optional) | #### Configuration Examples ```bash # Direct API usage (international endpoint) KIMI_API_KEY=your_kimi_api_key KIMI_SERVER_URL=https://api.moonshot.ai/v1 # Alternative endpoint KIMI_SERVER_URL=https://api.moonshot.cn/v1 # China # With LiteLLM proxy KIMI_API_KEY=your_litellm_key KIMI_SERVER_URL=http://litellm-proxy:4000 KIMI_PROVIDER=moonshot # Adds prefix to model names (moonshot/kimi-k2.5) for LiteLLM ``` #### Supported Models PentAGI supports 11 Kimi/Moonshot models with tool calling, streaming, thinking modes, and multimodal capabilities. Models marked with `*` are used in default configuration. **Kimi K2.5 Series - Advanced Multimodal** | Model ID | Thinking | Multimodal | Context | Speed | Price (Input/Output) | Use Case | | -------------------------- | -------- | ---------- | ------- | ---------- | -------------------- | ----------------------------------------------- | | `kimi-k2.5`* | ✅ | ✅ | 256K | Standard | $0.60/$3.00 | Most intelligent, versatile, vision+text+code | **Kimi K2 Series - MoE Foundation (1T params, 32B activated)** | Model ID | Thinking | Multimodal | Context | Speed | Price (Input/Output) | Use Case | | -------------------------- | -------- | ---------- | ------- | ---------- | -------------------- | ----------------------------------------------- | | `kimi-k2-0905-preview`* | ❌ | ❌ | 256K | Standard | $0.60/$2.50 | Enhanced agentic coding, improved frontend | | `kimi-k2-0711-preview` | ❌ | ❌ | 128K | Standard | $0.60/$2.50 | Powerful code and agent capabilities | | `kimi-k2-turbo-preview`* | ❌ | ❌ | 256K | Turbo | $1.15/$8.00 | High-speed version, 60-100 tokens/sec | | `kimi-k2-thinking` | ✅ | ❌ | 256K | Standard | $0.60/$2.50 | Long-term thinking, multi-step tool usage | | `kimi-k2-thinking-turbo` | ✅ | ❌ | 256K | Turbo | $1.15/$8.00 | High-speed thinking, deep reasoning | **Moonshot V1 Series - General Text Generation** | Model ID | Thinking | Multimodal | Context | Speed | Price (Input/Output) | Use Case | | -------------------------- | -------- | ---------- | ------- | ---------- | -------------------- | ----------------------------------------------- | | `moonshot-v1-8k` | ❌ | ❌ | 8K | Standard | $0.20/$2.00 | Short text generation, cost-effective | | `moonshot-v1-32k` | ❌ | ❌ | 32K | Standard | $1.00/$3.00 | Long text generation, balanced | | `moonshot-v1-128k` | ❌ | ❌ | 128K | Standard | $2.00/$5.00 | Very long text generation, extensive context | **Moonshot V1 Vision Series - Multimodal** | Model ID | Thinking | Multimodal | Context | Speed | Price (Input/Output) | Use Case | | ----------------------------- | -------- | ---------- | ------- | ---------- | -------------------- | ----------------------------------------------- | | `moonshot-v1-8k-vision-preview` | ❌ | ✅ | 8K | Standard | $0.20/$2.00 | Vision understanding, short context | | `moonshot-v1-32k-vision-preview` | ❌ | ✅ | 32K | Standard | $1.00/$3.00 | Vision understanding, medium context | | `moonshot-v1-128k-vision-preview` | ❌ | ✅ | 128K | Standard | $2.00/$5.00 | Vision understanding, long context | **Prices**: Per 1M tokens. Turbo models offer 60-100 tokens/sec output speed with higher pricing. **Key Features**: - **Ultra-Long Context**: Up to 256K tokens for comprehensive codebase analysis - **Multimodal Capabilities**: Vision models support image understanding for screenshot analysis (Kimi K2.5, V1 Vision series) - **Extended Thinking**: Deep reasoning with multi-step tool usage (kimi-k2.5, kimi-k2-thinking models) - **High-Speed Turbo**: 60-100 tokens/sec output for real-time workflows (Turbo variants) - **Tool Calling**: Seamless integration with 20+ pentesting tools via function calling - **Streaming**: Real-time response streaming for interactive security assessment - **Multilingual**: Strong Chinese and English language support - **MoE Architecture**: Efficient 1T total parameters with 32B activated for K2 series **LiteLLM Integration**: Set `KIMI_PROVIDER=moonshot` to enable model name prefixing when using default PentAGI configurations with LiteLLM proxy. Leave empty for direct API usage. ### Qwen Provider Configuration PentAGI integrates with Qwen from Alibaba Cloud Model Studio (DashScope), providing powerful multilingual models with reasoning capabilities and context caching support. #### Configuration Variables | Variable | Default Value | Description | | ------------------ | ------------------------------------------------------ | --------------------------------------------------- | | `QWEN_API_KEY` | | Qwen API key for authentication | | `QWEN_SERVER_URL` | `https://dashscope-us.aliyuncs.com/compatible-mode/v1` | Qwen API endpoint URL (international) | | `QWEN_PROVIDER` | | Provider prefix for LiteLLM integration (optional) | #### Configuration Examples ```bash # Direct API usage (Global/US endpoint) QWEN_API_KEY=your_qwen_api_key QWEN_SERVER_URL=https://dashscope-us.aliyuncs.com/compatible-mode/v1 # Alternative endpoints QWEN_SERVER_URL=https://dashscope-intl.aliyuncs.com/compatible-mode/v1 # International (Singapore) QWEN_SERVER_URL=https://dashscope.aliyuncs.com/compatible-mode/v1 # Chinese Mainland (Beijing) # With LiteLLM proxy QWEN_API_KEY=your_litellm_key QWEN_SERVER_URL=http://litellm-proxy:4000 QWEN_PROVIDER=dashscope # Adds prefix to model names (dashscope/qwen-plus) for LiteLLM ``` #### Supported Models PentAGI supports 32 Qwen models with tool calling, streaming, thinking modes, and context caching. Models marked with `*` are used in default configuration. **Wide Availability Models (All Regions)** | Model ID | Thinking | Intl | Global/US | China | Price (Input/Output/Cache) | Use Case | | ---------------------------- | -------- | ---- | --------- | ----- | -------------------------- | ----------------------------------------------- | | `qwen3-max`* | ✅ | ✅ | ✅ | ✅ | $2.40/$12.00/$0.48 | Flagship reasoning, complex security analysis | | `qwen3-max-preview` | ✅ | ✅ | ✅ | ✅ | $2.40/$12.00/$0.48 | Preview version with extended thinking | | `qwen-max` | ❌ | ✅ | ❌ | ✅ | $1.60/$6.40/$0.32 | Strong instruction following, legacy flagship | | `qwen3.5-plus`* | ✅ | ✅ | ✅ | ✅ | $0.40/$2.40/$0.08 | Balanced reasoning, general dialogue, coding | | `qwen-plus` | ✅ | ✅ | ✅ | ✅ | $0.40/$4.00/$0.08 | Cost-effective balanced performance | | `qwen3.5-flash`* | ✅ | ✅ | ✅ | ✅ | $0.10/$0.40/$0.02 | Ultra-fast lightweight, high-throughput | | `qwen-flash` | ❌ | ✅ | ✅ | ✅ | $0.05/$0.40/$0.01 | Fast with context caching, cost-optimized | | `qwen-turbo` | ✅ | ✅ | ❌ | ✅ | $0.05/$0.50/$0.01 | Deprecated, use qwen-flash instead | | `qwq-plus` | ✅ | ✅ | ❌ | ✅ | $0.80/$2.40/$0.16 | Deep reasoning, chain-of-thought analysis | **Region-Specific Models** | Model ID | Thinking | Intl | Global/US | China | Price (Input/Output/Cache) | Use Case | | ---------------------------- | -------- | ---- | --------- | ----- | -------------------------- | ----------------------------------------------- | | `qwen-plus-us` | ✅ | ❌ | ✅ | ❌ | $0.40/$4.00/$0.08 | US region optimized balanced model | | `qwen-long-latest` | ❌ | ❌ | ❌ | ✅ | $0.07/$0.29/$0.01 | Ultra-long context (10M tokens) | **Open Source - Qwen3.5 Series** | Model ID | Thinking | Intl | Global/US | China | Price (Input/Output/Cache) | Use Case | | ---------------------------- | -------- | ---- | --------- | ----- | -------------------------- | ----------------------------------------------- | | `qwen3.5-397b-a17b` | ✅ | ✅ | ✅ | ✅ | $0.60/$3.60/$0.12 | Largest 397B parameters, exceptional reasoning | | `qwen3.5-122b-a10b` | ✅ | ✅ | ✅ | ✅ | $0.40/$3.20/$0.08 | Large 122B parameters, strong performance | | `qwen3.5-27b` | ✅ | ✅ | ✅ | ✅ | $0.30/$2.40/$0.06 | Medium 27B parameters, balanced | | `qwen3.5-35b-a3b` | ✅ | ✅ | ✅ | ✅ | $0.25/$2.00/$0.05 | Efficient 35B with 3B active MoE | **Open Source - Qwen3 Series** | Model ID | Thinking | Intl | Global/US | China | Price (Input/Output/Cache) | Use Case | | ------------------------------ | -------- | ---- | --------- | ----- | -------------------------- | ----------------------------------------------- | | `qwen3-next-80b-a3b-thinking` | ✅ | ✅ | ✅ | ✅ | $0.15/$1.43/$0.03 | Next-gen 80B thinking-only mode | | `qwen3-next-80b-a3b-instruct` | ❌ | ✅ | ✅ | ✅ | $0.15/$1.20/$0.03 | Next-gen 80B instruction following | | `qwen3-235b-a22b` | ✅ | ✅ | ✅ | ✅ | $0.70/$8.40/$0.14 | Dual-mode 235B with 22B active | | `qwen3-32b` | ✅ | ✅ | ✅ | ✅ | $0.29/$2.87/$0.06 | Versatile 32B dual-mode | | `qwen3-30b-a3b` | ✅ | ✅ | ✅ | ✅ | $0.20/$2.40/$0.04 | Efficient 30B MoE architecture | | `qwen3-14b` | ✅ | ✅ | ✅ | ✅ | $0.35/$4.20/$0.07 | Medium 14B performance-cost balance | | `qwen3-8b` | ✅ | ✅ | ✅ | ✅ | $0.18/$2.10/$0.04 | Compact 8B efficiency optimized | | `qwen3-4b` | ✅ | ✅ | ❌ | ✅ | $0.11/$1.26/$0.02 | Lightweight 4B for simple tasks | | `qwen3-1.7b` | ✅ | ✅ | ❌ | ✅ | $0.11/$1.26/$0.02 | Ultra-compact 1.7B basic tasks | | `qwen3-0.6b` | ✅ | ✅ | ❌ | ✅ | $0.11/$1.26/$0.02 | Smallest 0.6B minimal resources | **Open Source - QwQ & Qwen2.5 Series** | Model ID | Thinking | Intl | Global/US | China | Price (Input/Output/Cache) | Use Case | | ---------------------------- | -------- | ---- | --------- | ----- | -------------------------- | ----------------------------------------------- | | `qwq-32b` | ✅ | ✅ | ✅ | ✅ | $0.29/$0.86/$0.06 | Open 32B reasoning, deep research | | `qwen2.5-14b-instruct-1m` | ❌ | ✅ | ❌ | ✅ | $0.81/$3.22/$0.16 | Extended 1M context, 14B parameters | | `qwen2.5-7b-instruct-1m` | ❌ | ✅ | ❌ | ✅ | $0.37/$1.47/$0.07 | Extended 1M context, 7B parameters | | `qwen2.5-72b-instruct` | ❌ | ✅ | ❌ | ✅ | $1.40/$5.60/$0.28 | Large 72B instruction following | | `qwen2.5-32b-instruct` | ❌ | ✅ | ❌ | ✅ | $0.70/$2.80/$0.14 | Medium 32B instruction following | | `qwen2.5-14b-instruct` | ❌ | ✅ | ❌ | ✅ | $0.35/$1.40/$0.07 | Compact 14B instruction following | | `qwen2.5-7b-instruct` | ❌ | ✅ | ❌ | ✅ | $0.18/$0.70/$0.04 | Small 7B instruction following | | `qwen2.5-3b-instruct` | ❌ | ❌ | ❌ | ✅ | $0.04/$0.13/$0.01 | Lightweight 3B Chinese Mainland only | **Prices**: Per 1M tokens. Cache pricing is for implicit context caching (20% of input cost). Models with thinking support include additional reasoning computation during CoT phase. **Region Availability**: - **Intl** (International): Singapore region (`dashscope-intl.aliyuncs.com`) - **Global/US**: US Virginia region (`dashscope-us.aliyuncs.com`) - **China**: Chinese Mainland Beijing region (`dashscope.aliyuncs.com`) **Key Features**: - **Automatic Context Caching**: 30-50% cost reduction on repeated context with implicit cache (20% of input price) - **Extended Thinking**: Chain-of-thought reasoning for complex security analysis (Qwen3-Max, QwQ, Qwen3.5-Plus) - **Tool Calling**: Seamless integration with 20+ pentesting tools via function calling - **Streaming**: Real-time response streaming for interactive workflows - **Multilingual**: Strong Chinese, English, and multi-language support - **Ultra-Long Context**: Up to 10M tokens with qwen-long-latest for massive codebase analysis **LiteLLM Integration**: Set `QWEN_PROVIDER=dashscope` to enable model name prefixing when using default PentAGI configurations with LiteLLM proxy. Leave empty for direct API usage. ## Advanced Setup ### Langfuse Integration Langfuse provides advanced capabilities for monitoring and analyzing AI agent operations. 1. Configure Langfuse environment variables in existing `.env` file.
Langfuse valuable environment variables ### Database Credentials - `LANGFUSE_POSTGRES_USER` and `LANGFUSE_POSTGRES_PASSWORD` - Langfuse PostgreSQL credentials - `LANGFUSE_CLICKHOUSE_USER` and `LANGFUSE_CLICKHOUSE_PASSWORD` - ClickHouse credentials - `LANGFUSE_REDIS_AUTH` - Redis password ### Encryption and Security Keys - `LANGFUSE_SALT` - Salt for hashing in Langfuse Web UI - `LANGFUSE_ENCRYPTION_KEY` - Encryption key (32 bytes in hex) - `LANGFUSE_NEXTAUTH_SECRET` - Secret key for NextAuth ### Admin Credentials - `LANGFUSE_INIT_USER_EMAIL` - Admin email - `LANGFUSE_INIT_USER_PASSWORD` - Admin password - `LANGFUSE_INIT_USER_NAME` - Admin username ### API Keys and Tokens - `LANGFUSE_INIT_PROJECT_PUBLIC_KEY` - Project public key (used from PentAGI side too) - `LANGFUSE_INIT_PROJECT_SECRET_KEY` - Project secret key (used from PentAGI side too) ### S3 Storage - `LANGFUSE_S3_ACCESS_KEY_ID` - S3 access key ID - `LANGFUSE_S3_SECRET_ACCESS_KEY` - S3 secret access key
2. Enable integration with Langfuse for PentAGI service in `.env` file. ```bash LANGFUSE_BASE_URL=http://langfuse-web:3000 LANGFUSE_PROJECT_ID= # default: value from ${LANGFUSE_INIT_PROJECT_ID} LANGFUSE_PUBLIC_KEY= # default: value from ${LANGFUSE_INIT_PROJECT_PUBLIC_KEY} LANGFUSE_SECRET_KEY= # default: value from ${LANGFUSE_INIT_PROJECT_SECRET_KEY} ``` 3. Run the Langfuse stack: ```bash curl -O https://raw.githubusercontent.com/vxcontrol/pentagi/master/docker-compose-langfuse.yml docker compose -f docker-compose.yml -f docker-compose-langfuse.yml up -d ``` Visit [localhost:4000](http://localhost:4000) to access Langfuse Web UI with credentials from `.env` file: - `LANGFUSE_INIT_USER_EMAIL` - Admin email - `LANGFUSE_INIT_USER_PASSWORD` - Admin password ### Monitoring and Observability For detailed system operation tracking, integration with monitoring tools is available. 1. Enable integration with OpenTelemetry and all observability services for PentAGI in `.env` file. ```bash OTEL_HOST=otelcol:8148 ``` 2. Run the observability stack: ```bash curl -O https://raw.githubusercontent.com/vxcontrol/pentagi/master/docker-compose-observability.yml docker compose -f docker-compose.yml -f docker-compose-observability.yml up -d ``` Visit [localhost:3000](http://localhost:3000) to access Grafana Web UI. > [!NOTE] > If you want to use Observability stack with Langfuse, you need to enable integration in `.env` file to set `LANGFUSE_OTEL_EXPORTER_OTLP_ENDPOINT` to `http://otelcol:4318`. > > To run all available stacks together (Langfuse, Graphiti, and Observability): > > ```bash > docker compose -f docker-compose.yml -f docker-compose-langfuse.yml -f docker-compose-graphiti.yml -f docker-compose-observability.yml up -d > ``` > > You can also register aliases for these commands in your shell to run it faster: > > ```bash > alias pentagi="docker compose -f docker-compose.yml -f docker-compose-langfuse.yml -f docker-compose-graphiti.yml -f docker-compose-observability.yml" > alias pentagi-up="docker compose -f docker-compose.yml -f docker-compose-langfuse.yml -f docker-compose-graphiti.yml -f docker-compose-observability.yml up -d" > alias pentagi-down="docker compose -f docker-compose.yml -f docker-compose-langfuse.yml -f docker-compose-graphiti.yml -f docker-compose-observability.yml down" > ``` ### Knowledge Graph Integration (Graphiti) PentAGI integrates with [Graphiti](https://github.com/vxcontrol/pentagi-graphiti), a temporal knowledge graph system powered by Neo4j, to provide advanced semantic understanding and relationship tracking for AI agent operations. The vxcontrol fork provides custom entity and edge types that are specific to pentesting purposes. #### What is Graphiti? Graphiti automatically extracts and stores structured knowledge from agent interactions, building a graph of entities, relationships, and temporal context. This enables: - **Semantic Memory**: Store and recall relationships between tools, targets, vulnerabilities, and techniques - **Contextual Understanding**: Track how different pentesting actions relate to each other over time - **Knowledge Reuse**: Learn from past penetration tests and apply insights to new assessments - **Advanced Querying**: Search for complex patterns like "What tools were effective against similar targets?" #### Enabling Graphiti The Graphiti knowledge graph is **optional** and disabled by default. To enable it: 1. Configure Graphiti environment variables in `.env` file: ```bash ## Graphiti knowledge graph settings GRAPHITI_ENABLED=true GRAPHITI_TIMEOUT=30 GRAPHITI_URL=http://graphiti:8000 GRAPHITI_MODEL_NAME=gpt-5-mini # Neo4j settings (used by Graphiti stack) NEO4J_USER=neo4j NEO4J_DATABASE=neo4j NEO4J_PASSWORD=devpassword NEO4J_URI=bolt://neo4j:7687 # OpenAI API key (required by Graphiti for entity extraction) OPEN_AI_KEY=your_openai_api_key ``` 2. Run the Graphiti stack along with the main PentAGI services: ```bash # Download the Graphiti compose file if needed curl -O https://raw.githubusercontent.com/vxcontrol/pentagi/master/docker-compose-graphiti.yml # Start PentAGI with Graphiti docker compose -f docker-compose.yml -f docker-compose-graphiti.yml up -d ``` 3. Verify Graphiti is running: ```bash # Check service health docker compose -f docker-compose.yml -f docker-compose-graphiti.yml ps graphiti neo4j # View Graphiti logs docker compose -f docker-compose.yml -f docker-compose-graphiti.yml logs -f graphiti # Access Neo4j Browser (optional) # Visit http://localhost:7474 and login with NEO4J_USER/NEO4J_PASSWORD # Access Graphiti API (optional, for debugging) # Visit http://localhost:8000/docs for Swagger API documentation ``` > [!NOTE] > The Graphiti service is defined in `docker-compose-graphiti.yml` as a separate stack. You must run both compose files together to enable the knowledge graph functionality. The pre-built Docker image `vxcontrol/graphiti:latest` is used by default. #### What Gets Stored When enabled, PentAGI automatically captures: - **Agent Responses**: All agent reasoning, analysis, and decisions - **Tool Executions**: Commands executed, tools used, and their results - **Context Information**: Flow, task, and subtask hierarchy ### GitHub and Google OAuth Integration OAuth integration with GitHub and Google allows users to authenticate using their existing accounts on these platforms. This provides several benefits: - Simplified login process without need to create separate credentials - Enhanced security through trusted identity providers - Access to user profile information from GitHub/Google accounts - Seamless integration with existing development workflows For using GitHub OAuth you need to create a new OAuth application in your GitHub account and set the `OAUTH_GITHUB_CLIENT_ID` and `OAUTH_GITHUB_CLIENT_SECRET` in `.env` file. For using Google OAuth you need to create a new OAuth application in your Google account and set the `OAUTH_GOOGLE_CLIENT_ID` and `OAUTH_GOOGLE_CLIENT_SECRET` in `.env` file. ### Docker Image Configuration PentAGI allows you to configure Docker image selection for executing various tasks. The system automatically chooses the most appropriate image based on the task type, but you can constrain this selection by specifying your preferred images: | Variable | Default | Description | | ---------------------------------- | ---------------------- | ----------------------------------------------------------- | | `DOCKER_DEFAULT_IMAGE` | `debian:latest` | Default Docker image for general tasks and ambiguous cases | | `DOCKER_DEFAULT_IMAGE_FOR_PENTEST` | `vxcontrol/kali-linux` | Default Docker image for security/penetration testing tasks | When these environment variables are set, AI agents will be limited to the image choices you specify. This is particularly useful for: - **Security Enforcement**: Restricting usage to only verified and trusted images - **Environment Standardization**: Using corporate or customized images across all operations - **Performance Optimization**: Utilizing pre-built images with necessary tools already installed Configuration examples: ```bash # Using a custom image for general tasks DOCKER_DEFAULT_IMAGE=mycompany/custom-debian:latest # Using a specialized image for penetration testing DOCKER_DEFAULT_IMAGE_FOR_PENTEST=mycompany/pentest-tools:v2.0 ``` > [!NOTE] > If a user explicitly specifies a particular Docker image in their task, the system will try to use that exact image, ignoring these settings. These variables only affect the system's automatic image selection process. ## Development ### Development Requirements - golang - nodejs - docker - postgres - commitlint ### Environment Setup #### Backend Setup Run once `cd backend && go mod download` to install needed packages. For generating swagger files have to run ```bash swag init -g ../../pkg/server/router.go -o pkg/server/docs/ --parseDependency --parseInternal --parseDepth 2 -d cmd/pentagi ``` before installing `swag` package via ```bash go install github.com/swaggo/swag/cmd/swag@v1.8.7 ``` For generating graphql resolver files have to run ```bash go run github.com/99designs/gqlgen --config ./gqlgen/gqlgen.yml ``` after that you can see the generated files in `pkg/graph` folder. For generating ORM methods (database package) from sqlc configuration ```bash docker run --rm -v $(pwd):/src -w /src --network pentagi-network -e DATABASE_URL="{URL}" sqlc/sqlc:1.27.0 generate -f sqlc/sqlc.yml ``` For generating Langfuse SDK from OpenAPI specification ```bash fern generate --local ``` and to install fern-cli ```bash npm install -g fern-api ``` #### Testing For running tests `cd backend && go test -v ./...` #### Frontend Setup Run once `cd frontend && npm install` to install needed packages. For generating graphql files have to run `npm run graphql:generate` which using `graphql-codegen.ts` file. Be sure that you have `graphql-codegen` installed globally: ```bash npm install -g graphql-codegen ``` After that you can run: * `npm run prettier` to check if your code is formatted correctly * `npm run prettier:fix` to fix it * `npm run lint` to check if your code is linted correctly * `npm run lint:fix` to fix it For generating SSL certificates you need to run `npm run ssl:generate` which using `generate-ssl.ts` file or it will be generated automatically when you run `npm run dev`. #### Backend Configuration Edit the configuration for `backend` in `.vscode/launch.json` file: - `DATABASE_URL` - PostgreSQL database URL (eg. `postgres://postgres:postgres@localhost:5432/pentagidb?sslmode=disable`) - `DOCKER_HOST` - Docker SDK API (eg. for macOS `DOCKER_HOST=unix:///Users//Library/Containers/com.docker.docker/Data/docker.raw.sock`) [more info](https://stackoverflow.com/a/62757128/5922857) Optional: - `SERVER_PORT` - Port to run the server (default: `8443`) - `SERVER_USE_SSL` - Enable SSL for the server (default: `false`) #### Frontend Configuration Edit the configuration for `frontend` in `.vscode/launch.json` file: - `VITE_API_URL` - Backend API URL. *Omit* the URL scheme (e.g., `localhost:8080` *NOT* `http://localhost:8080`) - `VITE_USE_HTTPS` - Enable SSL for the server (default: `false`) - `VITE_PORT` - Port to run the server (default: `8000`) - `VITE_HOST` - Host to run the server (default: `0.0.0.0`) ### Running the Application #### Backend Run the command(s) in `backend` folder: - Use `.env` file to set environment variables like a `source .env` - Run `go run cmd/pentagi/main.go` to start the server > [!NOTE] > The first run can take a while as dependencies and docker images need to be downloaded to setup the backend environment. #### Frontend Run the command(s) in `frontend` folder: - Run `npm install` to install the dependencies - Run `npm run dev` to run the web app - Run `npm run build` to build the web app Open your browser and visit the web app URL. ## Testing LLM Agents PentAGI includes a powerful utility called `ctester` for testing and validating LLM agent capabilities. This tool helps ensure your LLM provider configurations work correctly with different agent types, allowing you to optimize model selection for each specific agent role. The utility features parallel testing of multiple agents, detailed reporting, and flexible configuration options. ### Key Features - **Parallel Testing**: Tests multiple agents simultaneously for faster results - **Comprehensive Test Suite**: Evaluates basic completion, JSON responses, function calling, and penetration testing knowledge - **Detailed Reporting**: Generates markdown reports with success rates and performance metrics - **Flexible Configuration**: Test specific agents or test groups as needed - **Specialized Test Groups**: Includes domain-specific tests for cybersecurity and penetration testing scenarios ### Usage Scenarios #### For Developers (with local Go environment) If you've cloned the repository and have Go installed: ```bash # Default configuration with .env file cd backend go run cmd/ctester/*.go -verbose # Custom provider configuration go run cmd/ctester/*.go -config ../examples/configs/openrouter.provider.yml -verbose # Generate a report file go run cmd/ctester/*.go -config ../examples/configs/deepinfra.provider.yml -report ../test-report.md # Test specific agent types only go run cmd/ctester/*.go -agents simple,simple_json,primary_agent -verbose # Test specific test groups only go run cmd/ctester/*.go -groups basic,advanced -verbose ``` #### For Users (using Docker image) If you prefer to use the pre-built Docker image without setting up a development environment: ```bash # Using Docker to test with default environment docker run --rm -v $(pwd)/.env:/opt/pentagi/.env vxcontrol/pentagi /opt/pentagi/bin/ctester -verbose # Test with your custom provider configuration docker run --rm \ -v $(pwd)/.env:/opt/pentagi/.env \ -v $(pwd)/my-config.yml:/opt/pentagi/config.yml \ vxcontrol/pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/config.yml -agents simple,primary_agent,coder -verbose # Generate a detailed report docker run --rm \ -v $(pwd)/.env:/opt/pentagi/.env \ -v $(pwd):/opt/pentagi/output \ vxcontrol/pentagi /opt/pentagi/bin/ctester -report /opt/pentagi/output/report.md ``` #### Using Pre-configured Providers The Docker image comes with built-in support for major providers (OpenAI, Anthropic, Gemini, Ollama) and pre-configured provider files for additional services (OpenRouter, DeepInfra, DeepSeek, Moonshot, Novita): ```bash # Test with OpenRouter configuration docker exec -it pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/conf/openrouter.provider.yml # Test with DeepInfra configuration docker exec -it pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/conf/deepinfra.provider.yml # Test with DeepSeek configuration docker exec -it pentagi /opt/pentagi/bin/ctester -provider deepseek # Test with GLM configuration docker exec -it pentagi /opt/pentagi/bin/ctester -provider glm # Test with Kimi configuration docker exec -it pentagi /opt/pentagi/bin/ctester -provider kimi # Test with Qwen configuration docker exec -it pentagi /opt/pentagi/bin/ctester -provider qwen # Test with DeepSeek configuration file for custom provider docker exec -it pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/conf/deepseek.provider.yml # Test with Moonshot configuration file for custom provider docker exec -it pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/conf/moonshot.provider.yml # Test with Novita configuration docker exec -it pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/conf/novita.provider.yml # Test with OpenAI configuration docker exec -it pentagi /opt/pentagi/bin/ctester -type openai # Test with Anthropic configuration docker exec -it pentagi /opt/pentagi/bin/ctester -type anthropic # Test with Gemini configuration docker exec -it pentagi /opt/pentagi/bin/ctester -type gemini # Test with AWS Bedrock configuration docker exec -it pentagi /opt/pentagi/bin/ctester -type bedrock # Test with Custom OpenAI configuration docker exec -it pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/conf/custom-openai.provider.yml # Test with Ollama configuration (local inference) docker exec -it pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/conf/ollama-llama318b.provider.yml # Test with Ollama Qwen3 32B configuration (requires custom model creation) docker exec -it pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/conf/ollama-qwen332b-fp16-tc.provider.yml # Test with Ollama QwQ 32B configuration (requires custom model creation and 71.3GB VRAM) docker exec -it pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/conf/ollama-qwq32b-fp16-tc.provider.yml ``` To use these configurations, your `.env` file only needs to contain: ``` LLM_SERVER_URL=https://openrouter.ai/api/v1 # or https://api.deepinfra.com/v1/openai or https://api.openai.com/v1 or https://api.novita.ai/openai LLM_SERVER_KEY=your_api_key LLM_SERVER_MODEL= # Leave empty, as models are specified in the config LLM_SERVER_CONFIG_PATH=/opt/pentagi/conf/openrouter.provider.yml # or deepinfra.provider.ymll or custom-openai.provider.yml or novita.provider.yml LLM_SERVER_PROVIDER= # Provider name for LiteLLM proxy (e.g., openrouter, deepseek, moonshot, novita) LLM_SERVER_LEGACY_REASONING=false # Controls reasoning format, for OpenAI must be true (default: false) LLM_SERVER_PRESERVE_REASONING=false # Preserve reasoning content in multi-turn conversations (required by Moonshot, default: false) # For OpenAI (official API) OPEN_AI_KEY=your_openai_api_key # Your OpenAI API key OPEN_AI_SERVER_URL=https://api.openai.com/v1 # OpenAI API endpoint # For Anthropic (Claude models) ANTHROPIC_API_KEY=your_anthropic_api_key # Your Anthropic API key ANTHROPIC_SERVER_URL=https://api.anthropic.com/v1 # Anthropic API endpoint # For Gemini (Google AI) GEMINI_API_KEY=your_gemini_api_key # Your Google AI API key GEMINI_SERVER_URL=https://generativelanguage.googleapis.com # Google AI API endpoint # For AWS Bedrock (enterprise foundation models) BEDROCK_REGION=us-east-1 # AWS region for Bedrock service # Authentication (choose one method, priority: DefaultAuth > BearerToken > AccessKey): BEDROCK_DEFAULT_AUTH=false # Use AWS SDK credential chain (env vars, EC2 role, ~/.aws/credentials) BEDROCK_BEARER_TOKEN= # Bearer token authentication (takes priority over static credentials) BEDROCK_ACCESS_KEY_ID=your_aws_access_key # AWS access key ID (static credentials) BEDROCK_SECRET_ACCESS_KEY=your_aws_secret_key # AWS secret access key (static credentials) BEDROCK_SESSION_TOKEN= # AWS session token (optional, for temporary credentials with static auth) BEDROCK_SERVER_URL= # Optional custom Bedrock endpoint (VPC endpoints, local testing) # For Ollama (local server or cloud) OLLAMA_SERVER_URL= # Local: http://ollama-server:11434, Cloud: https://ollama.com OLLAMA_SERVER_API_KEY= # Required for Ollama Cloud (https://ollama.com/settings/keys), leave empty for local OLLAMA_SERVER_MODEL= OLLAMA_SERVER_CONFIG_PATH= OLLAMA_SERVER_PULL_MODELS_TIMEOUT= OLLAMA_SERVER_PULL_MODELS_ENABLED= OLLAMA_SERVER_LOAD_MODELS_ENABLED= # For DeepSeek (Chinese AI with strong reasoning) DEEPSEEK_API_KEY= # DeepSeek API key DEEPSEEK_SERVER_URL=https://api.deepseek.com # DeepSeek API endpoint DEEPSEEK_PROVIDER= # Optional: LiteLLM prefix (e.g., 'deepseek') # For GLM (Zhipu AI) GLM_API_KEY= # GLM API key GLM_SERVER_URL=https://api.z.ai/api/paas/v4 # GLM API endpoint (international) GLM_PROVIDER= # Optional: LiteLLM prefix (e.g., 'zai') # For Kimi (Moonshot AI) KIMI_API_KEY= # Kimi API key KIMI_SERVER_URL=https://api.moonshot.ai/v1 # Kimi API endpoint (international) KIMI_PROVIDER= # Optional: LiteLLM prefix (e.g., 'moonshot') # For Qwen (Alibaba Cloud DashScope) QWEN_API_KEY= # Qwen API key QWEN_SERVER_URL=https://dashscope-us.aliyuncs.com/compatible-mode/v1 # Qwen API endpoint (US) QWEN_PROVIDER= # Optional: LiteLLM prefix (e.g., 'dashscope') # For Ollama (local inference) use variables above OLLAMA_SERVER_URL=http://localhost:11434 OLLAMA_SERVER_MODEL=llama3.1:8b-instruct-q8_0 OLLAMA_SERVER_CONFIG_PATH=/opt/pentagi/conf/ollama-llama318b.provider.yml OLLAMA_SERVER_PULL_MODELS_ENABLED=false OLLAMA_SERVER_LOAD_MODELS_ENABLED=false ``` #### Using OpenAI with Unverified Organizations For OpenAI accounts with unverified organizations that don't have access to the latest reasoning models (o1, o3, o4-mini), you need to use a custom configuration. To use OpenAI with unverified organization accounts, configure your `.env` file as follows: ```bash LLM_SERVER_URL=https://api.openai.com/v1 LLM_SERVER_KEY=your_openai_api_key LLM_SERVER_MODEL= # Leave empty, models are specified in config LLM_SERVER_CONFIG_PATH=/opt/pentagi/conf/custom-openai.provider.yml LLM_SERVER_LEGACY_REASONING=true # Required for OpenAI reasoning format ``` This configuration uses the pre-built `custom-openai.provider.yml` file that maps all agent types to models available for unverified organizations, using `o3-mini` instead of models like `o1`, `o3`, and `o4-mini`. You can test this configuration using: ```bash # Test with custom OpenAI configuration for unverified accounts docker exec -it pentagi /opt/pentagi/bin/ctester -config /opt/pentagi/conf/custom-openai.provider.yml ``` > [!NOTE] > The `LLM_SERVER_LEGACY_REASONING=true` setting is crucial for OpenAI compatibility as it ensures reasoning parameters are sent in the format expected by OpenAI's API. #### Using LiteLLM Proxy When using LiteLLM proxy to access various LLM providers, model names are prefixed with the provider name (e.g., `moonshot/kimi-2.5` instead of `kimi-2.5`). To use the same provider configuration files with both direct API access and LiteLLM proxy, set the `LLM_SERVER_PROVIDER` variable: ```bash # Direct access to Moonshot API LLM_SERVER_URL=https://api.moonshot.ai/v1 LLM_SERVER_KEY=your_moonshot_api_key LLM_SERVER_CONFIG_PATH=/opt/pentagi/conf/moonshot.provider.yml LLM_SERVER_PROVIDER= # Empty for direct access # Access via LiteLLM proxy LLM_SERVER_URL=http://litellm-proxy:4000 LLM_SERVER_KEY=your_litellm_api_key LLM_SERVER_CONFIG_PATH=/opt/pentagi/conf/moonshot.provider.yml LLM_SERVER_PROVIDER=moonshot # Provider prefix for LiteLLM ``` With `LLM_SERVER_PROVIDER=moonshot`, the system automatically prefixes all model names from the configuration file with `moonshot/`, making them compatible with LiteLLM's model naming convention. **LiteLLM Provider Name Mapping:** When using LiteLLM proxy, set the corresponding `*_PROVIDER` variable to enable model prefixing: - `deepseek` - for DeepSeek models (`DEEPSEEK_PROVIDER=deepseek` → `deepseek/deepseek-chat`) - `zai` - for GLM models (`GLM_PROVIDER=zai` → `zai/glm-4`) - `moonshot` - for Kimi models (`KIMI_PROVIDER=moonshot` → `moonshot/kimi-k2.5`) - `dashscope` - for Qwen models (`QWEN_PROVIDER=dashscope` → `dashscope/qwen-plus`) - `openai`, `anthropic`, `gemini` - for major cloud providers - `openrouter` - for OpenRouter aggregator - `deepinfra` - for DeepInfra hosting - `novita` - for Novita AI - Any other provider name configured in your LiteLLM instance **Example with LiteLLM:** ```bash # Use DeepSeek models via LiteLLM proxy with model prefixing DEEPSEEK_API_KEY=your_litellm_proxy_key DEEPSEEK_SERVER_URL=http://litellm-proxy:4000 DEEPSEEK_PROVIDER=deepseek # Models become deepseek/deepseek-chat, deepseek/deepseek-reasoner for LiteLLM # Direct DeepSeek API usage (no prefix needed) DEEPSEEK_API_KEY=your_deepseek_api_key DEEPSEEK_SERVER_URL=https://api.deepseek.com # Leave DEEPSEEK_PROVIDER empty ``` This approach allows you to: - Use the same configuration files for both direct and proxied access - Switch between providers without modifying configuration files - Easily test different routing strategies with LiteLLM #### Running Tests in a Production Environment If you already have a running PentAGI container and want to test the current configuration: ```bash # Run ctester in an existing container using current environment variables docker exec -it pentagi /opt/pentagi/bin/ctester -verbose # Test specific agent types with deterministic ordering docker exec -it pentagi /opt/pentagi/bin/ctester -agents simple,primary_agent,pentester -groups basic,knowledge -verbose # Generate a report file inside the container docker exec -it pentagi /opt/pentagi/bin/ctester -report /opt/pentagi/data/agent-test-report.md # Access the report from the host docker cp pentagi:/opt/pentagi/data/agent-test-report.md ./ ``` ### Command-line Options The utility accepts several options: - `-env ` - Path to environment file (default: `.env`) - `-type ` - Provider type: `custom`, `openai`, `anthropic`, `ollama`, `bedrock`, `gemini` (default: `custom`) - `-config ` - Path to custom provider config (default: from `LLM_SERVER_CONFIG_PATH` env variable) - `-tests ` - Path to custom tests YAML file (optional) - `-report ` - Path to write the report file (optional) - `-agents ` - Comma-separated list of agent types to test (default: `all`) - `-groups ` - Comma-separated list of test groups to run (default: `all`) - `-verbose` - Enable verbose output with detailed test results for each agent ### Available Agent Types Agents are tested in the following deterministic order: 1. **simple** - Basic completion tasks 2. **simple_json** - JSON-structured responses 3. **primary_agent** - Main reasoning agent 4. **assistant** - Interactive assistant mode 5. **generator** - Content generation 6. **refiner** - Content refinement and improvement 7. **adviser** - Expert advice and consultation 8. **reflector** - Self-reflection and analysis 9. **searcher** - Information gathering and search 10. **enricher** - Data enrichment and expansion 11. **coder** - Code generation and analysis 12. **installer** - Installation and setup tasks 13. **pentester** - Penetration testing and security assessment ### Available Test Groups - **basic** - Fundamental completion and prompt response tests - **advanced** - Complex reasoning and function calling tests - **json** - JSON format validation and structure tests (specifically designed for `simple_json` agent) - **knowledge** - Domain-specific cybersecurity and penetration testing knowledge tests > **Note**: The `json` test group is specifically designed for the `simple_json` agent type, while all other agents are tested with `basic`, `advanced`, and `knowledge` groups. This specialization ensures optimal testing coverage for each agent's intended purpose. ### Example Provider Configuration Provider configuration defines which models to use for different agent types: ```yaml simple: model: "provider/model-name" temperature: 0.7 top_p: 0.95 n: 1 max_tokens: 4000 simple_json: model: "provider/model-name" temperature: 0.7 top_p: 1.0 n: 1 max_tokens: 4000 json: true # ... other agent types ... ``` ### Optimization Workflow 1. **Create a baseline**: Run tests with default configuration to establish benchmark performance 2. **Analyze agent-specific performance**: Review the deterministic agent ordering to identify underperforming agents 3. **Test specialized configurations**: Experiment with different models for each agent type using provider-specific configs 4. **Focus on domain knowledge**: Pay special attention to knowledge group tests for cybersecurity expertise 5. **Validate function calling**: Ensure tool-based tests pass consistently for critical agent types 6. **Compare results**: Look for the best success rate and performance across all test groups 7. **Deploy optimal configuration**: Use in production with your optimized setup This tool helps ensure your AI agents are using the most effective models for their specific tasks, improving reliability while optimizing costs. ## Embedding Configuration and Testing PentAGI uses vector embeddings for semantic search, knowledge storage, and memory management. The system supports multiple embedding providers that can be configured according to your needs and preferences. ### Supported Embedding Providers PentAGI supports the following embedding providers: - **OpenAI** (default): Uses OpenAI's text embedding models - **Ollama**: Local embedding model through Ollama - **Mistral**: Mistral AI's embedding models - **Jina**: Jina AI's embedding service - **HuggingFace**: Models from HuggingFace - **GoogleAI**: Google's embedding models - **VoyageAI**: VoyageAI's embedding models
Embedding Provider Configuration (click to expand) ### Environment Variables To configure the embedding provider, set the following environment variables in your `.env` file: ```bash # Primary embedding configuration EMBEDDING_PROVIDER=openai # Provider type (openai, ollama, mistral, jina, huggingface, googleai, voyageai) EMBEDDING_MODEL=text-embedding-3-small # Model name to use EMBEDDING_URL= # Optional custom API endpoint EMBEDDING_KEY= # API key for the provider (if required) EMBEDDING_BATCH_SIZE=100 # Number of documents to process in a batch EMBEDDING_STRIP_NEW_LINES=true # Whether to remove new lines from text before embedding # Advanced settings PROXY_URL= # Optional proxy for all API calls HTTP_CLIENT_TIMEOUT=600 # Timeout in seconds for external API calls (default: 600, 0 = no timeout) # SSL/TLS Certificate Configuration (for external communication with LLM backends and tool servers) EXTERNAL_SSL_CA_PATH= # Path to custom CA certificate file (PEM format) inside the container # Must point to /opt/pentagi/ssl/ directory (e.g., /opt/pentagi/ssl/ca-bundle.pem) EXTERNAL_SSL_INSECURE=false # Skip certificate verification (use only for testing) ```
How to Add Custom CA Certificates (click to expand) If you see this error: `tls: failed to verify certificate: x509: certificate signed by unknown authority` **Step 1:** Get your CA certificate bundle in PEM format (can contain multiple certificates) **Step 2:** Place the file in the SSL directory on your host machine: ```bash # Default location (if PENTAGI_SSL_DIR is not set) cp ca-bundle.pem ./pentagi-ssl/ # Or custom location (if using PENTAGI_SSL_DIR in docker-compose.yml) cp ca-bundle.pem /path/to/your/ssl/dir/ ``` **Step 3:** Set the path in `.env` file (path must be inside the container): ```bash # The volume pentagi-ssl is mounted to /opt/pentagi/ssl inside the container EXTERNAL_SSL_CA_PATH=/opt/pentagi/ssl/ca-bundle.pem EXTERNAL_SSL_INSECURE=false ``` **Step 4:** Restart PentAGI: ```bash docker compose restart pentagi ``` **Notes:** - The `pentagi-ssl` volume is mounted to `/opt/pentagi/ssl` inside the container - You can change host directory using `PENTAGI_SSL_DIR` variable in docker-compose.yml - File supports multiple certificates and intermediate CAs in one PEM file - Use `EXTERNAL_SSL_INSECURE=true` only for testing (not recommended for production)
### Provider-Specific Limitations Each provider has specific limitations and supported features: - **OpenAI**: Supports all configuration options - **Ollama**: Does not support `EMBEDDING_KEY` as it uses local models - **Mistral**: Does not support `EMBEDDING_MODEL` or custom HTTP client - **Jina**: Does not support custom HTTP client - **HuggingFace**: Requires `EMBEDDING_KEY` and supports all other options - **GoogleAI**: Does not support `EMBEDDING_URL`, requires `EMBEDDING_KEY` - **VoyageAI**: Supports all configuration options If `EMBEDDING_URL` and `EMBEDDING_KEY` are not specified, the system will attempt to use the corresponding LLM provider settings (e.g., `OPEN_AI_KEY` when `EMBEDDING_PROVIDER=openai`). ### Why Consistent Embedding Providers Matter It's crucial to use the same embedding provider consistently because: 1. **Vector Compatibility**: Different providers produce vectors with different dimensions and mathematical properties 2. **Semantic Consistency**: Changing providers can break semantic similarity between previously embedded documents 3. **Memory Corruption**: Mixed embeddings can lead to poor search results and broken knowledge base functionality If you change your embedding provider, you should flush and reindex your entire knowledge base (see `etester` utility below).
### Embedding Tester Utility (etester) PentAGI includes a specialized `etester` utility for testing, managing, and debugging embedding functionality. This tool is essential for diagnosing and resolving issues related to vector embeddings and knowledge storage.
Etester Commands (click to expand) ```bash # Test embedding provider and database connection cd backend go run cmd/etester/main.go test -verbose # Show statistics about the embedding database go run cmd/etester/main.go info # Delete all documents from the embedding database (use with caution!) go run cmd/etester/main.go flush # Recalculate embeddings for all documents (after changing provider) go run cmd/etester/main.go reindex # Search for documents in the embedding database go run cmd/etester/main.go search -query "How to install PostgreSQL" -limit 5 ``` ### Using Docker If you're running PentAGI in Docker, you can use etester from within the container: ```bash # Test embedding provider docker exec -it pentagi /opt/pentagi/bin/etester test # Show detailed database information docker exec -it pentagi /opt/pentagi/bin/etester info -verbose ``` ### Advanced Search Options The `search` command supports various filters to narrow down results: ```bash # Filter by document type docker exec -it pentagi /opt/pentagi/bin/etester search -query "Security vulnerability" -doc_type guide -threshold 0.8 # Filter by flow ID docker exec -it pentagi /opt/pentagi/bin/etester search -query "Code examples" -doc_type code -flow_id 42 # All available search options docker exec -it pentagi /opt/pentagi/bin/etester search -help ``` Available search parameters: - `-query STRING`: Search query text (required) - `-doc_type STRING`: Filter by document type (answer, memory, guide, code) - `-flow_id NUMBER`: Filter by flow ID (positive number) - `-answer_type STRING`: Filter by answer type (guide, vulnerability, code, tool, other) - `-guide_type STRING`: Filter by guide type (install, configure, use, pentest, development, other) - `-limit NUMBER`: Maximum number of results (default: 3) - `-threshold NUMBER`: Similarity threshold (0.0-1.0, default: 0.7) ### Common Troubleshooting Scenarios 1. **After changing embedding provider**: Always run `flush` or `reindex` to ensure consistency 2. **Poor search results**: Try adjusting the similarity threshold or check if embeddings are correctly generated 3. **Database connection issues**: Verify PostgreSQL is running with pgvector extension installed 4. **Missing API keys**: Check environment variables for your chosen embedding provider
## Function Testing with ftester PentAGI includes a versatile utility called `ftester` for debugging, testing, and developing specific functions and AI agent behaviors. While `ctester` focuses on testing LLM model capabilities, `ftester` allows you to directly invoke individual system functions and AI agent components with precise control over execution context. ### Key Features - **Direct Function Access**: Test individual functions without running the entire system - **Mock Mode**: Test functions without a live PentAGI deployment using built-in mocks - **Interactive Input**: Fill function arguments interactively for exploratory testing - **Detailed Output**: Color-coded terminal output with formatted responses and errors - **Context-Aware Testing**: Debug AI agents within the context of specific flows, tasks, and subtasks - **Observability Integration**: All function calls are logged to Langfuse and Observability stack ### Usage Modes #### Command Line Arguments Run ftester with specific function and arguments directly from the command line: ```bash # Basic usage with mock mode cd backend go run cmd/ftester/main.go [function_name] -[arg1] [value1] -[arg2] [value2] # Example: Test terminal command in mock mode go run cmd/ftester/main.go terminal -command "ls -la" -message "List files" # Using a real flow context go run cmd/ftester/main.go -flow 123 terminal -command "whoami" -message "Check user" # Testing AI agent in specific task/subtask context go run cmd/ftester/main.go -flow 123 -task 456 -subtask 789 pentester -message "Find vulnerabilities" ``` #### Interactive Mode Run ftester without arguments for a guided interactive experience: ```bash # Start interactive mode go run cmd/ftester/main.go [function_name] # For example, to interactively fill browser tool arguments go run cmd/ftester/main.go browser ```
Available Functions (click to expand) ### Environment Functions - **terminal**: Execute commands in a container and return the output - **file**: Perform file operations (read, write, list) in a container ### Search Functions - **browser**: Access websites and capture screenshots - **google**: Search the web using Google Custom Search - **duckduckgo**: Search the web using DuckDuckGo - **tavily**: Search using Tavily AI search engine - **traversaal**: Search using Traversaal AI search engine - **perplexity**: Search using Perplexity AI - **sploitus**: Search for security exploits, vulnerabilities (CVEs), and pentesting tools - **searxng**: Search using Searxng meta search engine (aggregates results from multiple engines) ### Vector Database Functions - **search_in_memory**: Search for information in vector database - **search_guide**: Find guidance documents in vector database - **search_answer**: Find answers to questions in vector database - **search_code**: Find code examples in vector database ### AI Agent Functions - **advice**: Get expert advice from an AI agent - **coder**: Request code generation or modification - **maintenance**: Run system maintenance tasks - **memorist**: Store and organize information in vector database - **pentester**: Perform security tests and vulnerability analysis - **search**: Complex search across multiple sources ### Utility Functions - **describe**: Show information about flows, tasks, and subtasks
Debugging Flow Context (click to expand) The `describe` function provides detailed information about tasks and subtasks within a flow. This is particularly useful for diagnosing issues when PentAGI encounters problems or gets stuck. ```bash # List all flows in the system go run cmd/ftester/main.go describe # Show all tasks and subtasks for a specific flow go run cmd/ftester/main.go -flow 123 describe # Show detailed information for a specific task go run cmd/ftester/main.go -flow 123 -task 456 describe # Show detailed information for a specific subtask go run cmd/ftester/main.go -flow 123 -task 456 -subtask 789 describe # Show verbose output with full descriptions and results go run cmd/ftester/main.go -flow 123 describe -verbose ``` This function allows you to identify the exact point where a flow might be stuck and resume processing by directly invoking the appropriate agent function.
Function Help and Discovery (click to expand) Each function has a help mode that shows available parameters: ```bash # Get help for a specific function go run cmd/ftester/main.go [function_name] -help # Examples: go run cmd/ftester/main.go terminal -help go run cmd/ftester/main.go browser -help go run cmd/ftester/main.go describe -help ``` You can also run ftester without arguments to see a list of all available functions: ```bash go run cmd/ftester/main.go ```
Output Format (click to expand) The `ftester` utility uses color-coded output to make interpretation easier: - **Blue headers**: Section titles and key names - **Cyan [INFO]**: General information messages - **Green [SUCCESS]**: Successful operations - **Red [ERROR]**: Error messages - **Yellow [WARNING]**: Warning messages - **Yellow [MOCK]**: Indicates mock mode operation - **Magenta values**: Function arguments and results JSON and Markdown responses are automatically formatted for readability.
Advanced Usage Scenarios (click to expand) ### Debugging Stuck AI Flows When PentAGI gets stuck in a flow: 1. Pause the flow through the UI 2. Use `describe` to identify the current task and subtask 3. Directly invoke the agent function with the same task/subtask IDs 4. Examine the detailed output to identify the issue 5. Resume the flow or manually intervene as needed ### Testing Environment Variables Verify that API keys and external services are configured correctly: ```bash # Test Google search API configuration go run cmd/ftester/main.go google -query "pentesting tools" # Test browser access to external websites go run cmd/ftester/main.go browser -url "https://example.com" ``` ### Developing New AI Agent Behaviors When developing new prompt templates or agent behaviors: 1. Create a test flow in the UI 2. Use ftester to directly invoke the agent with different prompts 3. Observe responses and adjust prompts accordingly 4. Check Langfuse for detailed traces of all function calls ### Verifying Docker Container Setup Ensure containers are properly configured: ```bash go run cmd/ftester/main.go -flow 123 terminal -command "env | grep -i proxy" -message "Check proxy settings" ```
Docker Container Usage (click to expand) If you have PentAGI running in Docker, you can use ftester from within the container: ```bash # Run ftester inside the running PentAGI container docker exec -it pentagi /opt/pentagi/bin/ftester [arguments] # Examples: docker exec -it pentagi /opt/pentagi/bin/ftester -flow 123 describe docker exec -it pentagi /opt/pentagi/bin/ftester -flow 123 terminal -command "ps aux" -message "List processes" ``` This is particularly useful for production deployments where you don't have a local development environment.
Integration with Observability Tools (click to expand) All function calls made through ftester are logged to: 1. **Langfuse**: Captures the entire AI agent interaction chain, including prompts, responses, and function calls 2. **OpenTelemetry**: Records metrics, traces, and logs for system performance analysis 3. **Terminal Output**: Provides immediate feedback on function execution To access detailed logs: - Check Langfuse UI for AI agent traces (typically at `http://localhost:4000`) - Use Grafana dashboards for system metrics (typically at `http://localhost:3000`) - Examine terminal output for immediate function results and errors
### Command-line Options The main utility accepts several options: - `-env ` - Path to environment file (optional, default: `.env`) - `-provider ` - Provider type to use (default: `custom`, options: `openai`, `anthropic`, `ollama`, `bedrock`, `gemini`, `custom`) - `-flow ` - Flow ID for testing (0 means using mocks, default: `0`) - `-task ` - Task ID for agent context (optional) - `-subtask ` - Subtask ID for agent context (optional) Function-specific arguments are passed after the function name using `-name value` format. ## Building ### Building Docker Image The Docker build process automatically embeds version information from git tags. To properly version your build, use the provided scripts: #### Linux/macOS ```bash # Load version variables source ./scripts/version.sh # Standard build docker build \ --build-arg PACKAGE_VER=$PACKAGE_VER \ --build-arg PACKAGE_REV=$PACKAGE_REV \ -t pentagi:$PACKAGE_VER . # Multi-platform build docker buildx build \ --platform linux/amd64,linux/arm64 \ --build-arg PACKAGE_VER=$PACKAGE_VER \ --build-arg PACKAGE_REV=$PACKAGE_REV \ -t pentagi:$PACKAGE_VER . # Build and push docker buildx build \ --platform linux/amd64,linux/arm64 \ --build-arg PACKAGE_VER=$PACKAGE_VER \ --build-arg PACKAGE_REV=$PACKAGE_REV \ -t myregistry/pentagi:$PACKAGE_VER \ --push . ``` #### Windows (PowerShell) ```powershell # Load version variables . .\scripts\version.ps1 # Standard build docker build ` --build-arg PACKAGE_VER=$env:PACKAGE_VER ` --build-arg PACKAGE_REV=$env:PACKAGE_REV ` -t pentagi:$env:PACKAGE_VER . # Multi-platform build docker buildx build ` --platform linux/amd64,linux/arm64 ` --build-arg PACKAGE_VER=$env:PACKAGE_VER ` --build-arg PACKAGE_REV=$env:PACKAGE_REV ` -t pentagi:$env:PACKAGE_VER . ``` #### Quick build without version For development builds without version tracking: ```bash docker build -t pentagi:dev . ``` > [!NOTE] > - The build scripts automatically determine version from git tags > - Release builds (on tag commit) have no revision suffix > - Development builds (after tag) include commit hash as revision (e.g., `1.1.0-bc6e800`) > - To use the built image locally, update the image name in `docker-compose.yml` or use the `build` option ## Credits This project is made possible thanks to the following research and developments: - [Emerging Architectures for LLM Applications](https://lilianweng.github.io/posts/2023-06-23-agent) - [A Survey of Autonomous LLM Agents](https://arxiv.org/abs/2403.08299) - [Codel](https://github.com/semanser/codel) by Andriy Semenets - initial architectural inspiration for agent-based automation ## License **PentAGI** is licensed under the [MIT License](LICENSE). Copyright (c) 2025 PentAGI Development Team ### Third-Party Dependencies All third-party dependencies use MIT-compatible licenses. See [licenses/](licenses/) directory for detailed license reports. ### VXControl Cloud Services ⚠️ **Note:** While the VXControl Cloud SDK code is MIT licensed, accessing **VXControl Cloud Services** (threat intelligence, AI support, premium features) requires a separate License Key and compliance with [Terms of Service](https://github.com/vxcontrol/cloud#license-and-terms). The SDK code itself is free to use - service access requires registration. For questions contact: **info@pentagi.com** or **info@vxcontrol.com**