- Add host network mode support in Docker client (DOCKER_NETWORK=host) - Update documentation for network modes (bridge vs host) - Enhance OOB port allocation guidance with mandatory directives - Improve terminal command execution descriptions (detach, timeout) - Fix MSF workflow issues: add process isolation rules and RPC daemon patterns - Add terminal execution mechanics to adviser prompts for better monitoring - Update installer locale with host network mode explanation Fixes agent issues with msfconsole hanging, port conflicts, and process isolation.
87 KiB
PentAGI Configuration Guide
This document serves as a comprehensive guide to the configuration system in PentAGI, primarily aimed at developers. It details all available configuration options, their purposes, default values, and how they're used throughout the application.
Table of Contents
- PentAGI Configuration Guide
- Table of Contents
- Configuration Basics
- General Settings
- Docker Settings
- Server Settings
- Frontend Settings
- Authentication Settings
- Web Scraper Settings
- LLM Provider Settings
- Embedding Settings
- Summarizer Settings
- Assistant Settings
- Functions Configuration
- Search Engine Settings
- Network and Proxy Settings
- Graphiti Knowledge Graph Settings
- Agent Supervision Settings
- Observability Settings
Configuration Basics
PentAGI uses environment variables for configuration, with support for .env files through the godotenv package. The configuration is defined in the Config struct in pkg/config/config.go and is loaded using the NewConfig() function.
func NewConfig() (*Config, error) {
godotenv.Load()
var config Config
if err := env.ParseWithOptions(&config, env.Options{
RequiredIfNoDef: false,
FuncMap: map[reflect.Type]env.ParserFunc{
reflect.TypeOf(&url.URL{}): func(s string) (interface{}, error) {
if s == "" {
return nil, nil
}
return url.Parse(s)
},
},
}); err != nil {
return nil, err
}
return &config, nil
}
This function automatically loads environment variables from a .env file if present, then parses them into the Config struct using the env package from github.com/caarlos0/env/v10.
General Settings
These settings control basic application behavior and are foundational for the system's operation.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| DatabaseURL | DATABASE_URL |
postgres://pentagiuser:pentagipass@pgvector:5432/pentagidb?sslmode=disable |
Connection string for the PostgreSQL database with pgvector extension |
| Debug | DEBUG |
false |
Enables debug mode with additional logging |
| DataDir | DATA_DIR |
./data |
Directory for storing persistent data |
| AskUser | ASK_USER |
false |
When enabled, requires explicit user confirmation for certain operations |
| InstallationID | INSTALLATION_ID |
(none) | Unique installation identifier for PentAGI Cloud API communication |
| LicenseKey | LICENSE_KEY |
(none) | License key for PentAGI Cloud API authentication and feature activation |
Usage Details
- DatabaseURL: This is a critical setting used throughout the application for all database connections. It's used to:
- Initialize the primary SQL database connection in
main.go - Create GORM ORM instances for model operations
- Configure pgvector connectivity for embedding operations
- Set up connection pools in various tools and executors
- Initialize the primary SQL database connection in
// In main.go for SQL connection
db, err := sql.Open("postgres", cfg.DatabaseURL)
// In main.go for GORM connection
orm, err := database.NewGorm(cfg.DatabaseURL, "postgres")
// In tools for vector database operations
pgvector.WithConnectionURL(fte.cfg.DatabaseURL)
- Debug: Controls debug mode throughout the application, enabling additional logging and development features:
- Activates detailed logging in the router setup
- Can enable development endpoints and tools
// In router.go for enabling debug mode
if cfg.Debug {
// Enable debug features
}
- DataDir: Specifies where PentAGI stores persistent data. This is used across multiple components:
- In
docker/client.gofor container volume mapping - For screenshots storage in
services.NewScreenshotService - In tools for file operations and data persistence
- In Docker container management for mapping volumes
- In
// In docker/client.go
dataDir, err := filepath.Abs(cfg.DataDir)
// In router.go for screenshot service
screenshotService := services.NewScreenshotService(orm, cfg.DataDir)
// In tools.go for various tools
dataDir: fte.cfg.DataDir
- AskUser: A safety feature that, when enabled, requires explicit user confirmation before executing potentially destructive operations:
- Used in tools to prompt for confirmation before executing commands
- Serves as a safeguard for sensitive operations
// In tools.go
if fte.cfg.AskUser {
// Prompt user for confirmation before executing
}
- InstallationID: A unique identifier for the PentAGI installation used for cloud API communication:
- Generated automatically during installation or can be manually set
- Required for certain cloud-based features and integrations
// Used in cloud SDK initialization
if cfg.InstallationID != "" {
// Initialize cloud API client with installation ID
}
- LicenseKey: Authentication key for PentAGI Cloud API and premium feature activation:
- Validates license and enables licensed features
- Required for enterprise features and support
- Used for authentication with PentAGI Cloud services
// Used in cloud SDK initialization
if cfg.LicenseKey != "" {
// Validate license and activate premium features
}
Docker Settings
These settings control how PentAGI interacts with Docker, which is used for terminal isolation and executing commands in a controlled environment. They're crucial for the security and functionality of tool execution.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| DockerInside | DOCKER_INSIDE |
false |
Set to true if PentAGI runs inside Docker and needs to access the host Docker daemon. |
| DockerNetAdmin | DOCKER_NET_ADMIN |
false |
Set to true to grant the primary container NET_ADMIN capability for advanced networking. |
| DockerSocket | DOCKER_SOCKET |
(none) | Path to Docker socket for container management |
| DockerNetwork | DOCKER_NETWORK |
(none) | Docker network name for bridge mode, or host for host network mode. See network modes below. |
| DockerPublicIP | DOCKER_PUBLIC_IP |
0.0.0.0 |
Public IP address for Docker containers' port bindings (bridge mode only) |
| DockerWorkDir | DOCKER_WORK_DIR |
(none) | Custom working directory inside Docker containers |
| DockerDefaultImage | DOCKER_DEFAULT_IMAGE |
debian:latest |
Default Docker image for containers when specific images fail |
| DockerDefaultImageForPentest | DOCKER_DEFAULT_IMAGE_FOR_PENTEST |
vxcontrol/kali-linux |
Default Docker image for penetration testing tasks |
Usage Details
The Docker settings are primarily used in pkg/docker/client.go which implements the Docker client interface used throughout the application. This client is responsible for creating, managing, and executing commands in Docker containers:
-
DockerInside: Signals whether PentAGI is running inside a Docker container itself, which affects how volumes and sockets are mounted:
inside := cfg.DockerInside -
DockerSocket: Specifies the path to the Docker socket, which is crucial for container management:
if cfg.DockerSocket != "" { socket = cfg.DockerSocket } -
DockerNetwork: Controls the network isolation mode for containers. Supports two modes:
Bridge Mode (custom network name, e.g.,
pentagi-network):- Containers run in an isolated bridge network
- Port forwarding maps container ports to host ports
- Enhanced security through network isolation
- Recommended for most deployments
Host Mode (special value:
host):- Containers share the host's network stack directly
- No port forwarding - services bind directly to host interfaces
- Required for advanced network testing (raw packets, custom protocols)
- Reduced isolation - use with caution
network := cfg.DockerNetwork // Host network mode if dc.network == "host" { hostConfig.NetworkMode = container.NetworkMode("host") // No port bindings needed } else if dc.network != "" { // Bridge mode with custom network networkingConfig = &network.NetworkingConfig{ EndpointsConfig: map[string]*network.EndpointSettings{ dc.network: {}, }, } // Port bindings are configured } -
DockerPublicIP: Defines the IP address to bind container ports to, making services accessible:
publicIP := cfg.DockerPublicIP // Used when setting up port bindings hostConfig.PortBindings[natPort] = []nat.PortBinding{ { HostIP: dc.publicIP, HostPort: fmt.Sprintf("%d", port), }, } -
DockerWorkDir: Provides a custom working directory path to use inside containers:
hostDir := getHostDataDir(ctx, cli, dataDir, cfg.DockerWorkDir) -
DockerDefaultImage: Specifies the fallback image to use when requested images aren't available:
defImage := strings.ToLower(cfg.DockerDefaultImage) if defImage == "" { defImage = defaultImage }
This client is used by the tools executor to run commands in isolated containers, providing a secure environment for AI agents to execute terminal commands.
Server Settings
These settings control the HTTP and GraphQL server that forms the backend API of PentAGI.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| ServerPort | SERVER_PORT |
8080 |
Port for the HTTP server |
| ServerHost | SERVER_HOST |
0.0.0.0 |
Host address for the HTTP server |
| ServerUseSSL | SERVER_USE_SSL |
false |
Enable SSL for the HTTP server |
| ServerSSLKey | SERVER_SSL_KEY |
(none) | Path to SSL key file |
| ServerSSLCrt | SERVER_SSL_CRT |
(none) | Path to SSL certificate file |
Usage Details
These settings are used in main.go to configure and start the HTTP server:
// Build the listen address from host and port
listen := net.JoinHostPort(cfg.ServerHost, strconv.Itoa(cfg.ServerPort))
// Conditionally use TLS based on SSL configuration
if cfg.ServerUseSSL && cfg.ServerSSLCrt != "" && cfg.ServerSSLKey != "" {
err = r.RunTLS(listen, cfg.ServerSSLCrt, cfg.ServerSSLKey)
} else {
err = r.Run(listen)
}
The settings determine:
- The IP address and port the server listens on
- Whether to use HTTPS (SSL/TLS) for secure connections
- The location of the SSL certificate and key files (when SSL is enabled)
These configurations are crucial for production deployments where proper server binding and secure communication are required.
Frontend Settings
These settings control how the server serves frontend assets and handles Cross-Origin Resource Sharing (CORS) for API requests from browsers.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| StaticURL | STATIC_URL |
(none) | URL to serve static frontend assets from (enables reverse proxy mode) |
| StaticDir | STATIC_DIR |
./fe |
Directory containing frontend static files (used when not in proxy mode) |
| CorsOrigins | CORS_ORIGINS |
* |
Allowed origins for CORS requests (comma-separated) |
Usage Details
The frontend settings are extensively used in pkg/server/router.go for configuring how the application serves the frontend:
-
StaticURL: When set, enables reverse proxy mode where static assets are served from an external URL:
if cfg.StaticURL != nil && cfg.StaticURL.Scheme != "" && cfg.StaticURL.Host != "" { // Set up reverse proxy for static assets router.NoRoute(func(c *gin.Context) { req := c.Request.Clone(c.Request.Context()) req.URL.Scheme = cfg.StaticURL.Scheme req.URL.Host = cfg.StaticURL.Host // ... }) } -
StaticDir: When StaticURL is not set, specifies the local directory containing static frontend assets:
// Serve static files from local directory router.Use(static.Serve("/", static.LocalFile(cfg.StaticDir, true))) // Also used for finding index.html for SPA routes indexPath := filepath.Join(cfg.StaticDir, "index.html") -
CorsOrigins: Configures CORS policy for the API, controlling which origins can make requests:
// In GraphQL service initialization graphqlService := services.NewGraphqlService(db, baseURL, cfg.CorsOrigins, providers, controller, subscriptions) // In CORS middleware configuration if !slices.Contains(cfg.CorsOrigins, "*") { config.AllowCredentials = true } config.AllowOrigins = cfg.CorsOrigins
These settings are essential for:
- Supporting different deployment architectures (single server vs. separate frontend/backend)
- Enabling proper SPA routing for frontend applications
- Configuring security policies for cross-origin requests
Authentication Settings
These settings control authentication mechanisms, including cookie-based sessions and OAuth providers for user login.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| CookieSigningSalt | COOKIE_SIGNING_SALT |
(none) | Salt for signing and securing cookies used in sessions |
| PublicURL | PUBLIC_URL |
(none) | Public URL for auth callbacks from OAuth providers |
| OAuthGoogleClientID | OAUTH_GOOGLE_CLIENT_ID |
(none) | Google OAuth client ID for authentication |
| OAuthGoogleClientSecret | OAUTH_GOOGLE_CLIENT_SECRET |
(none) | Google OAuth client secret |
| OAuthGithubClientID | OAUTH_GITHUB_CLIENT_ID |
(none) | GitHub OAuth client ID for authentication |
| OAuthGithubClientSecret | OAUTH_GITHUB_CLIENT_SECRET |
(none) | GitHub OAuth client secret |
Usage Details
The authentication settings are used in pkg/server/router.go to set up authentication middleware and OAuth providers:
-
CookieSigningSalt: Used to secure cookies for session management:
// Used in auth middleware for authentication checks authMiddleware := auth.NewAuthMiddleware(baseURL, cfg.CookieSigningSalt) // Used for cookie store creation cookieStore := cookie.NewStore(auth.MakeCookieStoreKey(cfg.CookieSigningSalt)...) router.Use(sessions.Sessions("auth", cookieStore)) -
PublicURL: The base URL for OAuth callback endpoints, crucial for redirects after authentication:
publicURL, err := url.Parse(cfg.PublicURL) -
OAuth Provider Settings: Used to configure authentication with Google and GitHub:
// Google OAuth setup if publicURL != nil && cfg.OAuthGoogleClientID != "" && cfg.OAuthGoogleClientSecret != "" { googleOAuth := oauth.NewGoogleOAuthController( cfg.OAuthGoogleClientID, cfg.OAuthGoogleClientSecret, *publicURL, ) // ... } // GitHub OAuth setup if publicURL != nil && cfg.OAuthGithubClientID != "" && cfg.OAuthGithubClientSecret != "" { githubOAuth := oauth.NewGithubOAuthController( cfg.OAuthGithubClientID, cfg.OAuthGithubClientSecret, *publicURL, ) // ... }
These settings are essential for:
- Secure user authentication and session management
- Supporting social login through OAuth providers
- Enabling proper redirects in the authentication flow
Web Scraper Settings
These settings control the web scraper service used for browsing websites and taking screenshots, which allows AI agents to interact with web content.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| ScraperPublicURL | SCRAPER_PUBLIC_URL |
(none) | Public URL for accessing the scraper service from clients |
| ScraperPrivateURL | SCRAPER_PRIVATE_URL |
(none) | Private URL for internal scraper service access |
Usage Details
The scraper settings are extensively used in the tools executor to provide web browsing capabilities to AI agents:
// In various tool functions in pkg/tools/tools.go
browseTool = &functions.BrowseFunc{
scPrvURL: fte.cfg.ScraperPrivateURL,
scPubURL: fte.cfg.ScraperPublicURL,
// ...
}
screenshotTool = &functions.ScreenshotFunc{
scPrvURL: fte.cfg.ScraperPrivateURL,
scPubURL: fte.cfg.ScraperPublicURL,
// ...
}
These URLs serve different purposes:
- ScraperPublicURL: Used when generating URLs that will be accessed by the client (browser)
- ScraperPrivateURL: Used for internal communication between the backend and the scraper service
The scraper settings enable critical functionality:
- Web browsing capabilities for AI agents
- Screenshot capturing for web content analysis
- Web information gathering for research tasks
LLM Provider Settings
These settings control the integration with various Large Language Model (LLM) providers, including OpenAI, Anthropic, and custom providers.
OpenAI
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| OpenAIKey | OPEN_AI_KEY |
(none) | API key for OpenAI services |
| OpenAIServerURL | OPEN_AI_SERVER_URL |
https://api.openai.com/v1 |
Server URL for OpenAI API requests |
Anthropic
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| AnthropicAPIKey | ANTHROPIC_API_KEY |
(none) | API key for Anthropic Claude services |
| AnthropicServerURL | ANTHROPIC_SERVER_URL |
https://api.anthropic.com/v1 |
Server URL for Anthropic API requests |
Ollama LLM Provider
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| OllamaServerURL | OLLAMA_SERVER_URL |
(none) | Ollama server URL (local or cloud https://ollama.com) |
| OllamaServerAPIKey | OLLAMA_SERVER_API_KEY |
(none) | Ollama Cloud API key (optional, required for https://ollama.com) |
| OllamaServerModel | OLLAMA_SERVER_MODEL |
(none) | Default model to use for inference |
| OllamaServerConfig | OLLAMA_SERVER_CONFIG_PATH |
(none) | Path to config file for Ollama provider options |
| OllamaServerPullModelsTimeout | OLLAMA_SERVER_PULL_MODELS_TIMEOUT |
600 |
Timeout in seconds for model downloads |
| OllamaServerPullModelsEnabled | OLLAMA_SERVER_PULL_MODELS_ENABLED |
false |
Automatically download required models on startup |
| OllamaServerLoadModelsEnabled | OLLAMA_SERVER_LOAD_MODELS_ENABLED |
false |
Load available models list from server API |
Deployment Scenarios:
- Local Server: Set
OLLAMA_SERVER_URLto local endpoint (e.g.,http://ollama-server:11434), leaveOLLAMA_SERVER_API_KEYempty - Ollama Cloud: Set
OLLAMA_SERVER_URL=https://ollama.comand provideOLLAMA_SERVER_API_KEYfrom https://ollama.com/settings/keys
Note: When OllamaServerLoadModelsEnabled=false, only the default model is available. Enable this to see all installed models in the UI.
Google AI (Gemini) LLM Provider
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| GeminiAPIKey | GEMINI_API_KEY |
(none) | API key for Google AI Gemini services |
| GeminiServerURL | GEMINI_SERVER_URL |
https://generativelanguage.googleapis.com |
Server URL for Gemini API requests |
AWS Bedrock LLM Provider
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| BedrockRegion | BEDROCK_REGION |
us-east-1 |
AWS region for Bedrock service |
| BedrockDefaultAuth | BEDROCK_DEFAULT_AUTH |
false |
Use default AWS SDK credential chain (environment variables, EC2 role, ~/.aws/credentials) - highest priority |
| BedrockBearerToken | BEDROCK_BEARER_TOKEN |
(none) | Bearer token for authentication - takes priority over static credentials |
| BedrockAccessKey | BEDROCK_ACCESS_KEY_ID |
(none) | AWS access key ID for static credentials authentication |
| BedrockSecretKey | BEDROCK_SECRET_ACCESS_KEY |
(none) | AWS secret access key for static credentials authentication |
| BedrockSessionToken | BEDROCK_SESSION_TOKEN |
(none) | AWS session token for temporary credentials (optional, used with static credentials for STS/assumed roles) |
| BedrockServerURL | BEDROCK_SERVER_URL |
(none) | Optional custom endpoint URL for Bedrock service (VPC endpoints, local testing) |
Authentication Priority: BedrockDefaultAuth (highest) → BedrockBearerToken → BedrockAccessKey+BedrockSecretKey (lowest)
DeepSeek LLM Provider
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| DeepSeekAPIKey | DEEPSEEK_API_KEY |
(none) | DeepSeek API key for authentication |
| DeepSeekServerURL | DEEPSEEK_SERVER_URL |
https://api.deepseek.com |
DeepSeek API endpoint URL |
| DeepSeekProvider | DEEPSEEK_PROVIDER |
(none) | Provider name prefix for LiteLLM integration (optional) |
LiteLLM Integration: Set DEEPSEEK_PROVIDER=deepseek to enable model prefixing (e.g., deepseek/deepseek-chat) when using LiteLLM proxy with default PentAGI configs.
GLM LLM Provider
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| GLMAPIKey | GLM_API_KEY |
(none) | GLM API key for authentication |
| GLMServerURL | GLM_SERVER_URL |
https://api.z.ai/api/paas/v4 |
GLM API endpoint URL (international) |
| GLMProvider | GLM_PROVIDER |
(none) | Provider name prefix for LiteLLM integration (optional) |
Alternative Endpoints:
- International:
https://api.z.ai/api/paas/v4(default) - China:
https://open.bigmodel.cn/api/paas/v4 - Coding-specific:
https://api.z.ai/api/coding/paas/v4
LiteLLM Integration: Set GLM_PROVIDER=zai to enable model prefixing (e.g., zai/glm-4) when using LiteLLM proxy with default PentAGI configs.
Kimi LLM Provider
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| KimiAPIKey | KIMI_API_KEY |
(none) | Kimi API key for authentication |
| KimiServerURL | KIMI_SERVER_URL |
https://api.moonshot.ai/v1 |
Kimi API endpoint URL (international) |
| KimiProvider | KIMI_PROVIDER |
(none) | Provider name prefix for LiteLLM integration (optional) |
Alternative Endpoints:
- International:
https://api.moonshot.ai/v1(default) - China:
https://api.moonshot.cn/v1
LiteLLM Integration: Set KIMI_PROVIDER=moonshot to enable model prefixing (e.g., moonshot/kimi-k2.5) when using LiteLLM proxy with default PentAGI configs.
Qwen LLM Provider
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| QwenAPIKey | QWEN_API_KEY |
(none) | Qwen API key for authentication |
| QwenServerURL | QWEN_SERVER_URL |
https://dashscope-us.aliyuncs.com/compatible-mode/v1 |
Qwen API endpoint URL (international) |
| QwenProvider | QWEN_PROVIDER |
(none) | Provider name prefix for LiteLLM integration (optional) |
Alternative Endpoints:
- US:
https://dashscope-us.aliyuncs.com/compatible-mode/v1(default) - Singapore:
https://dashscope-intl.aliyuncs.com/compatible-mode/v1 - China:
https://dashscope.aliyuncs.com/compatible-mode/v1
LiteLLM Integration: Set QWEN_PROVIDER=dashscope to enable model prefixing (e.g., dashscope/qwen-plus) when using LiteLLM proxy with default PentAGI configs.
Custom LLM Provider
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| LLMServerURL | LLM_SERVER_URL |
(none) | Server URL for custom LLM provider |
| LLMServerKey | LLM_SERVER_KEY |
(none) | API key for custom LLM provider |
| LLMServerModel | LLM_SERVER_MODEL |
(none) | Model name for custom LLM provider |
| LLMServerConfig | LLM_SERVER_CONFIG_PATH |
(none) | Path to config file for custom LLM provider options |
| LLMServerProvider | LLM_SERVER_PROVIDER |
(none) | Provider name prefix for model names (useful for LiteLLM proxy) |
| LLMServerLegacyReasoning | LLM_SERVER_LEGACY_REASONING |
false |
Controls reasoning format in API requests |
| LLMServerPreserveReasoning | LLM_SERVER_PRESERVE_REASONING |
false |
Preserve reasoning content in multi-turn conversations (required by some providers) |
Usage Details
The LLM provider settings are used in pkg/providers modules to initialize and configure the appropriate language model providers:
-
OpenAI Settings: Used in
pkg/providers/openai/openai.goto create the OpenAI client:baseURL := cfg.OpenAIServerURL client, err := openai.New( openai.WithToken(cfg.OpenAIKey), openai.WithModel(OpenAIAgentModel), openai.WithBaseURL(baseURL), // ... ) -
Anthropic Settings: Used in
pkg/providers/anthropic/anthropic.goto create the Anthropic client:baseURL := cfg.AnthropicServerURL client, err := anthropic.New( anthropic.WithToken(cfg.AnthropicAPIKey), anthropic.WithBaseURL(baseURL), // ... ) -
Ollama Settings: Used in
pkg/providers/ollama/ollama.goto create the Ollama client:serverURL := cfg.OllamaServerURL client, err := ollama.New( ollama.WithServerURL(serverURL), ollama.WithHTTPClient(httpClient), ollama.WithModel(OllamaAgentModel), ollama.WithPullModel(), ) // Load provider options from config file if specified if cfg.OllamaServerConfig != "" { configData, err := os.ReadFile(cfg.OllamaServerConfig) providerConfig, err := BuildProviderConfig(cfg, configData) // ... } -
Gemini Settings: Used in
pkg/providers/gemini/gemini.goto create the Google AI client:opts := []googleai.Option{ googleai.WithRest(), googleai.WithAPIKey(cfg.GeminiAPIKey), googleai.WithEndpoint(cfg.GeminiServerURL), googleai.WithDefaultModel(GeminiAgentModel), } client, err := googleai.New(context.Background(), opts...) -
Bedrock Settings: Used in
pkg/providers/bedrock/bedrock.goto create the AWS Bedrock client:opts := []func(*bconfig.LoadOptions) error{ bconfig.WithRegion(cfg.BedrockRegion), bconfig.WithCredentialsProvider(credentials.NewStaticCredentialsProvider( cfg.BedrockAccessKey, cfg.BedrockSecretKey, cfg.BedrockSessionToken, )), } if cfg.BedrockServerURL != "" { opts = append(opts, bconfig.WithBaseEndpoint(cfg.BedrockServerURL)) } bcfg, err := bconfig.LoadDefaultConfig(context.Background(), opts...) bclient := bedrockruntime.NewFromConfig(bcfg) client, err := bedrock.New( bedrock.WithClient(bclient), bedrock.WithModel(BedrockAgentModel), bedrock.WithConverseAPI(), )The
BedrockSessionTokenis optional and only required when using temporary AWS credentials (e.g., from STS, assumed roles, or MFA-enabled IAM users). For permanent IAM user credentials, leave this field empty. -
Custom LLM Settings: Used in
pkg/providers/custom/custom.goto create a custom LLM client:baseKey := cfg.LLMServerKey baseURL := cfg.LLMServerURL baseModel := cfg.LLMServerModel client, err := openai.New( openai.WithToken(baseKey), openai.WithModel(baseModel), openai.WithBaseURL(baseURL), // ... ) // Load provider options from config file if specified if cfg.LLMServerConfig != "" { providerConfig, err := LoadConfig(cfg.LLMServerConfig, simple) // ... } -
LLMServerLegacyReasoning: Controls the reasoning format used in API requests to custom LLM providers:
// Used in custom provider to determine reasoning format if cfg.LLMServerLegacyReasoning { // Uses legacy string-based reasoning_effort parameter } else { // Uses modern structured reasoning object with max_tokens }false(default): Uses modern format where reasoning is sent as a structured object withmax_tokensparametertrue: Uses legacy format with string-basedreasoning_effortparameter
This setting is important when working with different LLM providers as they may expect different reasoning formats in their API requests. If you encounter reasoning-related errors with custom providers, try changing this setting.
- LLMServerPreserveReasoning: Controls whether reasoning content is preserved and sent back in multi-turn conversations:
// Used in custom provider to preserve reasoning content if cfg.LLMServerPreserveReasoning { // Preserves and returns reasoning_content in assistant messages }false(default): Reasoning content is not preserved in conversation historytrue: Reasoning content is preserved and sent in subsequent API calls
This setting is required by some LLM providers (e.g., Moonshot) that return errors like "thinking is enabled but reasoning_content is missing in assistant tool call message" when reasoning content is not included in multi-turn conversations. Enable this setting if your provider requires reasoning content to be preserved across conversation turns.
The provider registration is managed in pkg/providers/providers.go:
// Provider registration based on available credentials
if cfg.OpenAIKey != "" {
p, err := openai.New(cfg, defaultConfigs[provider.ProviderOpenAI])
if err != nil {
return nil, fmt.Errorf("failed to create openai provider: %w", err)
}
providers[provider.DefaultProviderNameOpenAI] = p
}
if cfg.AnthropicAPIKey != "" {
p, err := anthropic.New(cfg, defaultConfigs[provider.ProviderAnthropic])
if err != nil {
return nil, fmt.Errorf("failed to create anthropic provider: %w", err)
}
providers[provider.DefaultProviderNameAnthropic] = p
}
if cfg.GeminiAPIKey != "" {
p, err := gemini.New(cfg, defaultConfigs[provider.ProviderGemini])
if err != nil {
return nil, fmt.Errorf("failed to create gemini provider: %w", err)
}
providers[provider.DefaultProviderNameGemini] = p
}
if cfg.BedrockAccessKey != "" && cfg.BedrockSecretKey != "" {
p, err := bedrock.New(cfg, defaultConfigs[provider.ProviderBedrock])
if err != nil {
return nil, fmt.Errorf("failed to create bedrock provider: %w", err)
}
providers[provider.DefaultProviderNameBedrock] = p
}
if cfg.OllamaServerURL != "" {
p, err := ollama.New(cfg, defaultConfigs[provider.ProviderOllama])
if err != nil {
return nil, fmt.Errorf("failed to create ollama provider: %w", err)
}
providers[provider.DefaultProviderNameOllama] = p
}
if cfg.LLMServerURL != "" && (cfg.LLMServerModel != "" || cfg.LLMServerConfig != "") {
p, err := custom.New(cfg, defaultConfigs[provider.ProviderCustom])
if err != nil {
return nil, fmt.Errorf("failed to create custom provider: %w", err)
}
providers[provider.DefaultProviderNameCustom] = p
}
These settings are critical for:
- Connecting to various LLM providers for AI capabilities
- Supporting multiple model options for different tasks
- Enabling custom or self-hosted LLM solutions
- Configuring specific model behaviors and parameters
Embedding Settings
These settings control the vector embedding service used for semantic search and similarity matching, which is fundamental for PentAGI's intelligent search capabilities.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| EmbeddingURL | EMBEDDING_URL |
(none) | Server URL for embedding provider (overrides provider-specific URLs) |
| EmbeddingKey | EMBEDDING_KEY |
(none) | API key for embedding provider (overrides provider-specific keys) |
| EmbeddingModel | EMBEDDING_MODEL |
(none) | Model name for embedding generation |
| EmbeddingStripNewLines | EMBEDDING_STRIP_NEW_LINES |
true |
Whether to strip newlines before embedding (improves quality) |
| EmbeddingBatchSize | EMBEDDING_BATCH_SIZE |
512 |
Batch size for embedding operations (affects memory usage and performance) |
| EmbeddingProvider | EMBEDDING_PROVIDER |
openai |
Provider for embeddings (openai, ollama, mistral, jina, huggingface) |
Usage Details
The embedding settings are extensively used in pkg/providers/embeddings/embedder.go to configure the vector embedding service:
-
EmbeddingProvider: Determines which embedding provider to use:
switch cfg.EmbeddingProvider { case "openai": return newOpenAIEmbedder(ctx, cfg) case "ollama": return newOllamaEmbedder(ctx, cfg) case "mistral": return newMistralEmbedder(ctx, cfg) case "jina": return newJinaEmbedder(ctx, cfg) case "huggingface": return newHuggingFaceEmbedder(ctx, cfg) default: return &embedder{nil}, fmt.Errorf("unsupported embedding provider: %s", cfg.EmbeddingProvider) } -
Provider-specific configurations: Used to configure each embedding provider with appropriate options:
// Example for OpenAI embeddings if cfg.EmbeddingURL != "" { opts = append(opts, openai.WithBaseURL(cfg.EmbeddingURL)) } else if cfg.OpenAIServerURL != "" { opts = append(opts, openai.WithBaseURL(cfg.OpenAIServerURL)) } if cfg.EmbeddingKey != "" { opts = append(opts, openai.WithToken(cfg.EmbeddingKey)) } else if cfg.OpenAIKey != "" { opts = append(opts, openai.WithToken(cfg.OpenAIKey)) } if cfg.EmbeddingModel != "" { opts = append(opts, openai.WithEmbeddingModel(cfg.EmbeddingModel)) } -
Embedding behavior configuration: Controls how text is processed for embeddings:
embeddings.WithStripNewLines(cfg.EmbeddingStripNewLines), embeddings.WithBatchSize(cfg.EmbeddingBatchSize),
These settings are essential for:
- Configuring semantic search capabilities
- Determining which embedding model to use
- Optimizing embedding performance and quality
- Supporting multiple embedding providers for flexibility
Summarizer Settings
These settings control the text summarization behavior used for condensing long conversations and improving context management in AI interactions. The summarization system is a critical component that allows PentAGI to maintain coherent, long-running conversations while managing token usage effectively.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| SummarizerPreserveLast | SUMMARIZER_PRESERVE_LAST |
true |
Preserve the last message in summarization |
| SummarizerUseQA | SUMMARIZER_USE_QA |
true |
Use question-answer format for summarization |
| SummarizerSumHumanInQA | SUMMARIZER_SUM_MSG_HUMAN_IN_QA |
false |
Include human messages in QA summaries |
| SummarizerLastSecBytes | SUMMARIZER_LAST_SEC_BYTES |
51200 |
Bytes to preserve from the last section (50KB) |
| SummarizerMaxBPBytes | SUMMARIZER_MAX_BP_BYTES |
16384 |
Maximum bytes for bullet points summarization (16KB) |
| SummarizerMaxQASections | SUMMARIZER_MAX_QA_SECTIONS |
10 |
Maximum QA sections to include |
| SummarizerMaxQABytes | SUMMARIZER_MAX_QA_BYTES |
65536 |
Maximum bytes for QA summarization (64KB) |
| SummarizerKeepQASections | SUMMARIZER_KEEP_QA_SECTIONS |
1 |
Number of recent QA sections to keep without summarization |
Usage Details and Impact on System Behavior
The summarizer settings map directly to the SummarizerConfig structure that controls the chain summarization algorithm in pkg/csum. These settings work together to implement a sophisticated, multi-strategy approach to managing conversation context:
Core Summarization Strategies and Their Parameters
-
Section Summarization - Always active, ensures all older sections (except the last one) consist of a single summarized pair
- No specific parameters control this as it's a fundamental part of the algorithm
- Prevents unbounded growth by consolidating completed conversation sections
-
Last Section Management (
SummarizerPreserveLastandSummarizerLastSecBytes)- Controls how the current/active conversation section is managed
- When
SummarizerPreserveLast = true, older messages within the last section will be summarized when the section exceedsSummarizerLastSecBytesbytes - A reserve space of 25% is automatically maintained to accommodate new messages without triggering frequent re-summarization
- Individual oversized pairs are summarized separately if they exceed
SummarizerMaxBPBytes
-
QA Pair Summarization (
SummarizerUseQA,SummarizerMaxQASections,SummarizerMaxQABytes,SummarizerSumHumanInQA)- When
SummarizerUseQA = true, creates larger summarization units focused on question-answer patterns - Preserves the most recent
SummarizerMaxQASectionssections as long as they don't exceedSummarizerMaxQABytestotal - If
SummarizerSumHumanInQA = true, human messages are also summarized; otherwise, they're preserved verbatim
- When
Deep Dive: Parameter Impact and Recommendations
SummarizerPreserveLast (Default: true)
- Purpose: Controls whether the last (active) section has size management applied
- Impact: When enabled, prevents the active conversation from growing indefinitely
- When to adjust:
- Enable (default) for production systems and long-running conversations
- Disable only for debugging or when you need to preserve the complete conversation history regardless of size
SummarizerLastSecBytes (Default: 51200 - 50KB)
- Purpose: Maximum byte size for the last (active) section before summarization begins
- Impact: Directly controls how much conversation history is preserved verbatim in the active section
- When to adjust:
- Increase for models with larger context windows to maintain more conversation detail
- Decrease for models with smaller context to prevent token limits from being exceeded
- Balance with
SummarizerMaxBPBytesto ensure coherent summarization
SummarizerMaxBPBytes (Default: 16384 - 16KB)
- Purpose: Maximum byte size for individual body pairs (typically AI responses)
- Impact: Controls when individual large responses get summarized, even if the overall section is under limit
- When to adjust:
- Increase if your use case involves long but important AI responses that shouldn't be summarized
- Decrease if you want more aggressive summarization of lengthy responses
SummarizerUseQA (Default: true)
- Purpose: Enables question-answer style summarization that creates more cohesive summaries
- Impact: When enabled, creates a new first section with a summary of older interactions, preserving recent sections
- When to adjust:
- Enable (default) for more coherent, organized summaries focused on main topics
- Disable if you prefer simpler, section-by-section summarization without cross-section analysis
SummarizerMaxQASections (Default: 10)
- Purpose: Maximum number of recent sections to preserve when using QA-style summarization
- Impact: Directly controls how many conversation turns remain intact after QA summarization
- When to adjust:
- Increase to preserve more recent conversation context (at the cost of token usage)
- Decrease to create more compact conversation histories, focusing on only the very recent exchanges
SummarizerMaxQABytes (Default: 65536 - 64KB)
- Purpose: Maximum total byte size for preserved sections in QA-style summarization
- Impact: Sets an upper bound on memory used by preserved sections, regardless of section count
- When to adjust:
- Increase for models with larger context windows or when detailed context is essential
- Decrease for smaller context models or when prioritizing efficiency over context preservation
SummarizerSumHumanInQA (Default: false)
- Purpose: Controls whether human messages are summarized in QA-style summarization
- Impact: When false, human messages are preserved verbatim; when true, they are also summarized
- When to adjust:
- Keep disabled (default) to preserve the exact wording of user queries
- Enable only when human messages are very verbose and token efficiency is critical
SummarizerKeepQASections (Default: 1)
- Purpose: Controls the number of recent QA sections to keep without summarization
- Impact: Directly controls how many recent conversation turns are preserved verbatim
- When to adjust:
- Increase to preserve more recent conversation context
- Decrease to create more compact conversation histories, focusing on only the very recent exchanges
Summarization Effects on Agent Behavior
The summarization settings have significant effects on agent behavior:
-
Context Retention vs. Token Efficiency
- More aggressive summarization (smaller byte limits) reduces token usage but may lose context details
- More permissive settings (larger byte limits) preserve more context but increase token consumption
-
Conversation Coherence
- Appropriate summarization helps the agent maintain a coherent understanding of the conversation
- Over-aggressive summarization may cause the agent to lose important details or previous instructions
- Under-aggressive summarization may lead to context overflow in longer conversations
-
Response Quality
- QA-style summarization (
SummarizerUseQA = true) typically improves response quality for complex tasks - Preserving human messages (
SummarizerSumHumanInQA = false) helps maintain alignment with user intent - Appropriate
SummarizerMaxBPBytesprevents loss of detailed information from complex AI responses
- QA-style summarization (
Implementation Details
The summarizer settings are used in pkg/providers/providers.go to configure the summarization behavior:
summarizer := provider.SummarizerSettings{
PreserveLast: cfg.SummarizerPreserveLast,
UseQA: cfg.SummarizerUseQA,
SummHumanInQA: cfg.SummarizerSumHumanInQA,
LastSecBytes: cfg.SummarizerLastSecBytes,
MaxBPBytes: cfg.SummarizerMaxBPBytes,
MaxQASections: cfg.SummarizerMaxQASections,
MaxQABytes: cfg.SummarizerMaxQABytes,
}
These settings are passed to various components through the chain summarization system:
// In csum/chain_summary.go
func NewSummarizer(config SummarizerConfig) Summarizer {
if config.PreserveLast {
if config.LastSecBytes <= 0 {
config.LastSecBytes = maxLastSectionByteSize
}
}
if config.UseQA {
if config.MaxQASections <= 0 {
config.MaxQASections = maxQAPairSections
}
if config.MaxQABytes <= 0 {
config.MaxQABytes = maxQAPairByteSize
}
}
if config.MaxBPBytes <= 0 {
config.MaxBPBytes = maxSingleBodyPairByteSize
}
return &summarizer{config: config}
}
Recommended Settings for Different Use Cases
-
Long-running Assistant Conversations
SummarizerPreserveLast: true SummarizerLastSecBytes: 51200 (50KB) SummarizerMaxBPBytes: 16384 (16KB) SummarizerUseQA: true SummarizerMaxQASections: 10 SummarizerMaxQABytes: 65536 (64KB) SummarizerSumHumanInQA: false SummarizerKeepQASections: 1The default settings are optimized for assistant-style conversations. They maintain a good balance between context retention and token efficiency.
-
Technical Problem-Solving with Large Context Models
SummarizerPreserveLast: true SummarizerLastSecBytes: 81920 (80KB) SummarizerMaxBPBytes: 32768 (32KB) SummarizerUseQA: true SummarizerMaxQASections: 15 SummarizerMaxQABytes: 102400 (100KB) SummarizerSumHumanInQA: false SummarizerKeepQASections: 1Increased limits to preserve more technical details when using models with large context windows (e.g., GPT-4).
-
Limited Context Models
SummarizerPreserveLast: true SummarizerLastSecBytes: 25600 (25KB) SummarizerMaxBPBytes: 8192 (8KB) SummarizerUseQA: true SummarizerMaxQASections: 5 SummarizerMaxQABytes: 32768 (32KB) SummarizerSumHumanInQA: true SummarizerKeepQASections: 1More aggressive summarization for models with smaller context windows (e.g., smaller or older LLMs).
-
Debugging or Analysis (Maximum Context Preservation)
SummarizerPreserveLast: false SummarizerUseQA: false SummarizerKeepQASections: 0Disables active summarization to preserve the complete conversation history for debugging purposes. Note that this can lead to context overflow in long conversations.
Assistant Settings
These settings control the behavior of the AI assistant functionality, including whether to use multi-agent delegation and assistant-specific summarization settings.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| AssistantUseAgents | ASSISTANT_USE_AGENTS |
false |
Controls the default value for agent usage when creating new assistants |
| AssistantSummarizerPreserveLast | ASSISTANT_SUMMARIZER_PRESERVE_LAST |
true |
Whether to preserve all messages in the assistant's last section |
| AssistantSummarizerLastSecBytes | ASSISTANT_SUMMARIZER_LAST_SEC_BYTES |
76800 |
Maximum byte size for assistant's last section (75KB) |
| AssistantSummarizerMaxBPBytes | ASSISTANT_SUMMARIZER_MAX_BP_BYTES |
16384 |
Maximum byte size for a single body pair in assistant context (16KB) |
| AssistantSummarizerMaxQASections | ASSISTANT_SUMMARIZER_MAX_QA_SECTIONS |
7 |
Maximum QA sections to preserve in assistant context |
| AssistantSummarizerMaxQABytes | ASSISTANT_SUMMARIZER_MAX_QA_BYTES |
76800 |
Maximum byte size for assistant's QA sections (75KB) |
| AssistantSummarizerKeepQASections | ASSISTANT_SUMMARIZER_KEEP_QA_SECTIONS |
3 |
Number of recent QA sections to preserve without summarization |
Usage Details
The assistant settings are used to configure the behavior of the AI assistant and its context management:
-
AssistantUseAgents: Controls the default state of the "Use Agents" toggle when creating new assistants in the UI:
// This setting affects the initial state when creating assistants // Users can always override this by toggling the "Use Agents" button in the UIfalse(default): New assistants are created with agent delegation disabled by defaulttrue: New assistants are created with agent delegation enabled by default
-
Assistant Summarizer Settings: These provide dedicated summarization configuration for assistant instances, typically allowing for more memory retention compared to the global settings:
// Assistant summarizer configuration provides more context retention // compared to global settings, preserving more recent conversation history // while still ensuring efficient token usage
The assistant summarizer configuration is designed to provide more memory for context retention compared to the global settings, preserving more recent conversation history while still ensuring efficient token usage.
Recommended Assistant Settings for Different Use Cases
-
Standard Assistant Conversations
AssistantUseAgents: false AssistantSummarizerPreserveLast: true AssistantSummarizerLastSecBytes: 76800 (75KB) AssistantSummarizerMaxBPBytes: 16384 (16KB) AssistantSummarizerMaxQASections: 7 AssistantSummarizerMaxQABytes: 76800 (75KB) AssistantSummarizerKeepQASections: 3The default settings provide a balance between context retention and performance for typical assistant interactions.
-
Multi-Agent Assistant Workflows
AssistantUseAgents: true AssistantSummarizerPreserveLast: true AssistantSummarizerLastSecBytes: 102400 (100KB) AssistantSummarizerMaxBPBytes: 32768 (32KB) AssistantSummarizerMaxQASections: 10 AssistantSummarizerMaxQABytes: 102400 (100KB) AssistantSummarizerKeepQASections: 5Enhanced settings for complex workflows that benefit from agent delegation with increased context preservation.
-
Resource-Constrained Assistant
AssistantUseAgents: false AssistantSummarizerPreserveLast: true AssistantSummarizerLastSecBytes: 51200 (50KB) AssistantSummarizerMaxBPBytes: 16384 (16KB) AssistantSummarizerMaxQASections: 5 AssistantSummarizerMaxQABytes: 51200 (50KB) AssistantSummarizerKeepQASections: 2More conservative settings for environments with limited resources or smaller context models.
Functions Configuration
These settings control which tools are available to AI agents and allow adding custom external functions. The Functions API enables fine-grained control over agent capabilities by selectively disabling built-in tools or extending functionality with custom integrations.
| Field | Type | Description |
|---|---|---|
| token | string (optional) | API token for authenticating external function calls |
| disabled | DisableFunction[] | List of built-in functions to disable for specific agent types |
| functions | ExternalFunction[] | List of custom external functions to add to agent capabilities |
DisableFunction Structure
Allows disabling specific built-in functions for certain agent contexts, providing security and control over agent capabilities.
| Field | Type | Description |
|---|---|---|
| name | string | Name of the built-in function to disable (e.g., terminal, browser, file) |
| context | string[] | Agent contexts where the function should be disabled (optional) |
Available Agent Contexts: agent, adviser, coder, searcher, generator, memorist, enricher, reporter, assistant
When context is empty or omitted, the function is disabled for all agents.
ExternalFunction Structure
Allows adding custom external functions that agents can call via HTTP endpoints, enabling integration with external tools and services.
| Field | Type | Description |
|---|---|---|
| name | string | Name of the custom function (must be unique) |
| url | string | HTTP(S) URL endpoint for the function |
| timeout | int64 | Timeout in seconds for function execution (default: 60) |
| context | string[] | Agent contexts where the function is available (optional) |
| schema | Schema object | JSON schema defining function parameters and description (OpenAI format) |
Available Agent Contexts: Same as DisableFunction (agent, adviser, coder, searcher, generator, memorist, enricher, reporter, assistant)
When context is empty or omitted, the function is available to all agents.
Usage Details
The Functions configuration is typically provided when creating a flow through the API:
// Example from pkg/tools/tools.go
type Functions struct {
Token *string `json:"token,omitempty"`
Disabled []DisableFunction `json:"disabled,omitempty"`
Function []ExternalFunction `json:"functions,omitempty"`
}
These settings are used in pkg/tools/tools.go to configure available tools for each agent type:
-
Token: Used for authenticating requests to external function endpoints:
// The token is passed in the Authorization header when calling external functions req.Header.Set("Authorization", "Bearer " + *functions.Token) -
Disabled: Filters out built-in functions for specific agent contexts:
// Check if function is disabled for current agent context for _, disabled := range functions.Disabled { if disabled.Name == functionName && (len(disabled.Context) == 0 || contains(disabled.Context, agentType)) { // Skip this function } } -
Functions: Adds custom external functions to agent capabilities:
// Register external functions as available tools for _, externalFunc := range functions.Function { if len(externalFunc.Context) == 0 || contains(externalFunc.Context, agentType) { definitions = append(definitions, externalFunc.Schema) handlers[externalFunc.Name] = createExternalHandler(externalFunc) } }
Example Configuration
{
"token": "secret-api-token-for-external-functions",
"disabled": [
{
"name": "terminal",
"context": ["searcher", "enricher"]
},
{
"name": "browser",
"context": ["memorist"]
},
{
"name": "file"
}
],
"functions": [
{
"name": "custom_vulnerability_scan",
"url": "https://scanner.example.com/api/v1/scan",
"timeout": 120,
"context": ["pentester", "coder"],
"schema": {
"type": "function",
"function": {
"name": "custom_vulnerability_scan",
"description": "Perform a custom vulnerability scan on the target",
"parameters": {
"type": "object",
"properties": {
"target": {
"type": "string",
"description": "Target IP address or domain to scan"
},
"scan_type": {
"type": "string",
"enum": ["quick", "full", "stealth"],
"description": "Type of scan to perform"
}
},
"required": ["target"]
}
}
}
},
{
"name": "query_threat_intelligence",
"url": "https://threatintel.example.com/api/query",
"timeout": 30,
"context": ["searcher", "adviser"],
"schema": {
"type": "function",
"function": {
"name": "query_threat_intelligence",
"description": "Query threat intelligence database for IoCs and TTPs",
"parameters": {
"type": "object",
"properties": {
"indicator": {
"type": "string",
"description": "IP, domain, hash, or other indicator to search"
},
"indicator_type": {
"type": "string",
"enum": ["ip", "domain", "hash", "url"],
"description": "Type of indicator"
}
},
"required": ["indicator", "indicator_type"]
}
}
}
}
]
}
Security Considerations
- Token Security: Store the
tokenvalue securely and use HTTPS endpoints for external functions - Function Validation: External functions should validate all inputs and return structured error messages
- Timeout Configuration: Set appropriate timeouts to prevent long-running operations from blocking agents
- Context Restriction: Use the
contextfield to limit which agents can access sensitive functions - URL Validation: Ensure external function URLs are trusted and properly secured
Built-in Functions Reference
Common built-in functions that can be disabled:
terminal- Execute shell commands in containersfile- Read and write files in containersbrowser- Browse websites and take screenshotssearch_in_memory- Search vector memory storesearch_guide- Search knowledge guidessearch_answer- Search for answerssearch_code- Search code repositoriesstore_guide- Store knowledge guidesstore_answer- Store answersstore_code- Store code snippetsgoogle- Google Searchduckduckgo- DuckDuckGo Searchtavily- Tavily Searchtraversaal- Traversaal Searchperplexity- Perplexity Searchsearxng- SearXNG Searchsploitus- Sploitus Exploit Searchgraphiti_search- Graphiti Knowledge Graph Search
The specific functions available depend on the agent type and system configuration.
Search Engine Settings
These settings control the integration with various search engines used for web search capabilities, providing AI agents with up-to-date information from the internet.
DuckDuckGo Search
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| DuckDuckGoEnabled | DUCKDUCKGO_ENABLED |
true |
Enable or disable DuckDuckGo Search engine |
| DuckDuckGoRegion | DUCKDUCKGO_REGION |
(none) | Region code for search results (e.g., us-en, uk-en, cn-zh) |
| DuckDuckGoSafeSearch | DUCKDUCKGO_SAFESEARCH |
(none) | Safe search filter (off, moderate, strict) |
| DuckDuckGoTimeRange | DUCKDUCKGO_TIME_RANGE |
(none) | Time range for search results (d: day, w: week, m: month, y: year) |
Sploitus Search
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| SploitusEnabled | SPLOITUS_ENABLED |
true |
Enable or disable Sploitus exploit and vulnerability search |
Google Search
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| GoogleAPIKey | GOOGLE_API_KEY |
(none) | API key for Google Search |
| GoogleCXKey | GOOGLE_CX_KEY |
(none) | Custom Search Engine ID for Google Search |
| GoogleLRKey | GOOGLE_LR_KEY |
lang_en |
Language restriction for Google Search (e.g., lang_en) |
Traversaal Search
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| TraversaalAPIKey | TRAVERSAAL_API_KEY |
(none) | API key for Traversaal search engine |
Tavily Search
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| TavilyAPIKey | TAVILY_API_KEY |
(none) | API key for Tavily search engine |
Perplexity Search
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| PerplexityAPIKey | PERPLEXITY_API_KEY |
(none) | API key for Perplexity search engine |
| PerplexityModel | PERPLEXITY_MODEL |
sonar |
Model to use for Perplexity search |
| PerplexityContextSize | PERPLEXITY_CONTEXT_SIZE |
low |
Context size for Perplexity search (low, medium, high) |
Searxng Search
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| SearxngURL | SEARXNG_URL |
(none) | Base URL for Searxng meta search engine instance |
| SearxngCategories | SEARXNG_CATEGORIES |
general |
Search categories to use (e.g., general, news, web) |
| SearxngLanguage | SEARXNG_LANGUAGE |
(none) | Language filter for search results (e.g., en, ch) |
| SearxngSafeSearch | SEARXNG_SAFESEARCH |
0 |
Safe search filter level (0 = none, 1 = moderate, 2 = strict) |
| SearxngTimeRange | SEARXNG_TIME_RANGE |
(none) | Time range filter (e.g., day, month, year) |
| SearxngTimeout | SEARXNG_TIMEOUT |
(none) | Request timeout in seconds for Searxng API calls |
Usage Details
The search engine settings are used in pkg/tools/tools.go to configure various search providers that AI agents can use:
// Google Search configuration
googleSearch: &functions.GoogleSearchFunc{
apiKey: fte.cfg.GoogleAPIKey,
cxKey: fte.cfg.GoogleCXKey,
lrKey: fte.cfg.GoogleLRKey,
proxyURL: fte.cfg.ProxyURL,
},
// Traversaal Search configuration
traversaalSearch: &functions.TraversaalSearchFunc{
apiKey: fte.cfg.TraversaalAPIKey,
proxyURL: fte.cfg.ProxyURL,
},
// Tavily Search configuration
tavilySearch: &functions.TavilySearchFunc{
apiKey: fte.cfg.TavilyAPIKey,
proxyURL: fte.cfg.ProxyURL,
summarizer: cfg.Summarizer,
},
// Perplexity Search configuration
perplexitySearch: &functions.PerplexitySearchFunc{
apiKey: fte.cfg.PerplexityAPIKey,
proxyURL: fte.cfg.ProxyURL,
model: fte.cfg.PerplexityModel,
contextSize: fte.cfg.PerplexityContextSize,
summarizer: cfg.Summarizer,
},
// Sploitus Search configuration
sploitus := NewSploitusTool(
fte.flowID,
cfg.TaskID,
cfg.SubtaskID,
fte.cfg.SploitusEnabled,
fte.cfg.ProxyURL,
fte.slp,
)
// Searxng Search configuration
searxng := NewSearxngTool(
fte.flowID,
cfg.TaskID,
cfg.SubtaskID,
fte.cfg.SearxngURL,
fte.cfg.SearxngCategories,
fte.cfg.SearxngLanguage,
fte.cfg.SearxngSafeSearch,
fte.cfg.SearxngTimeRange,
fte.cfg.ProxyURL,
fte.cfg.SearxngTimeout,
fte.slp,
cfg.Summarizer,
)
These settings enable:
- Access to multiple search engines for diverse information sources
- Configuration of search parameters like language, context size, and time range
- Integration of search capabilities into the AI agent's toolset
- Web information gathering with different search strategies
- Security research through Sploitus, providing access to exploit databases and CVE information
- Meta-search capabilities through Searxng, aggregating results from multiple search engines
Having multiple search engine options ensures redundancy and provides different search algorithms for varied information needs. Sploitus is specifically designed for security research, providing comprehensive exploit and vulnerability information essential for penetration testing. Searxng is particularly useful as it provides aggregated results from multiple search engines while offering enhanced privacy and customization options.
Network and Proxy Settings
These settings control HTTP proxy, SSL configuration, and network timeouts for outbound connections, which are important for network security and access control.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| ProxyURL | PROXY_URL |
(none) | URL for HTTP proxy (e.g., http://user:pass@proxy:8080) |
| ExternalSSLCAPath | EXTERNAL_SSL_CA_PATH |
(none) | Path to trusted CA certificate for external LLM SSL connections |
| ExternalSSLInsecure | EXTERNAL_SSL_INSECURE |
false |
Skip SSL certificate verification for external connections |
| HTTPClientTimeout | HTTP_CLIENT_TIMEOUT |
600 |
Timeout in seconds for external API calls (0 = no timeout) |
Usage Details
The proxy settings are used in various places to configure HTTP clients for external API calls:
// Example from openai.go, anthropic.go, and other provider files
if cfg.ProxyURL != "" {
httpClient = &http.Client{
Transport: &http.Transport{
Proxy: func(req *http.Request) (*url.URL, error) {
return url.Parse(cfg.ProxyURL)
},
},
}
}
The proxy URL is also passed to various tools that make external requests:
// In tools.go for search tools
googleSearch: &functions.GoogleSearchFunc{
apiKey: fte.cfg.GoogleAPIKey,
cxKey: fte.cfg.GoogleCXKey,
lrKey: fte.cfg.GoogleLRKey,
proxyURL: fte.cfg.ProxyURL,
},
The proxy setting is essential for:
- Routing all outbound API requests through a controlled proxy
- Implementing network-level security policies
- Enabling access to external services from restricted networks
- Monitoring and auditing external API usage
The SSL settings provide additional security configuration:
-
ExternalSSLCAPath: Specifies a custom CA certificate for validating SSL connections to external services:
// Used in provider initialization to configure custom CA certificates if cfg.ExternalSSLCAPath != "" { caCert, err := os.ReadFile(cfg.ExternalSSLCAPath) // Configure TLS with custom CA }This is useful when connecting to LLM providers with self-signed certificates or internal CAs.
-
ExternalSSLInsecure: Allows skipping SSL certificate verification:
// Used in HTTP client configuration if cfg.ExternalSSLInsecure { tlsConfig.InsecureSkipVerify = true }Warning: Only use this in development or trusted environments. Skipping certificate verification exposes connections to man-in-the-middle attacks.
-
HTTPClientTimeout: Sets the timeout for all external HTTP requests (LLM providers, search engines, etc.):
// Used in pkg/system/utils.go for HTTP client configuration timeout := defaultHTTPClientTimeout if cfg.HTTPClientTimeout > 0 { timeout = time.Duration(cfg.HTTPClientTimeout) * time.Second } httpClient := &http.Client{ Timeout: timeout, }The default value of 600 seconds (10 minutes) is suitable for most LLM API calls, including long-running operations. Setting this to 0 disables the timeout (not recommended in production), while very low values may cause legitimate requests to fail. This setting affects:
- All LLM provider API calls (OpenAI, Anthropic, Bedrock, etc.)
- Search engine requests (Google, Tavily, Perplexity, etc.)
- External tool integrations
- Embedding generation requests
Adjust this value based on your network conditions and the complexity of operations being performed.
Graphiti Knowledge Graph Settings
These settings control the integration with Graphiti, a temporal knowledge graph system powered by Neo4j, for advanced semantic understanding and relationship tracking of AI agent operations.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| GraphitiEnabled | GRAPHITI_ENABLED |
false |
Enable or disable Graphiti knowledge graph integration |
| GraphitiURL | GRAPHITI_URL |
http://localhost:8001 |
Base URL for Graphiti API service |
| GraphitiTimeout | GRAPHITI_TIMEOUT |
30 |
Timeout in seconds for Graphiti operations |
Usage Details
The Graphiti settings are used in pkg/graphiti/client.go and integrated throughout the provider system to automatically capture agent interactions and tool executions:
-
GraphitiEnabled: Controls whether the knowledge graph integration is active:
// Check if Graphiti is enabled if !cfg.GraphitiEnabled { return &Client{enabled: false}, nil } -
GraphitiURL: Specifies the Graphiti API endpoint:
client := graphiti.NewClient(cfg.GraphitiURL, timeout, cfg.GraphitiEnabled) -
GraphitiTimeout: Sets the maximum time for knowledge graph operations:
timeout := time.Duration(cfg.GraphitiTimeout) * time.Second storeCtx, cancel := context.WithTimeout(ctx, timeout) defer cancel()
The Graphiti integration captures:
- Agent responses and reasoning for all agent types (pentester, researcher, coder, etc.)
- Tool execution details including function name, arguments, results, and execution status
- Context information including flow, task, and subtask IDs for hierarchical organization
- Temporal relationships between entities, actions, and outcomes
These settings enable:
- Building a comprehensive knowledge base from agent interactions
- Semantic memory across multiple penetration tests
- Advanced querying of relationships between tools, targets, and techniques
- Learning from past successful approaches and strategies
The integration is designed to be non-blocking - if Graphiti operations fail, they are logged but don't interrupt the agent workflow.
Agent Supervision Settings
These settings control the agent supervision system, including execution monitoring and tool call limits for different agent types.
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| ExecutionMonitorEnabled | EXECUTION_MONITOR_ENABLED |
false |
Enable automatic execution monitoring (mentor/adviser supervision) |
| ExecutionMonitorSameToolLimit | EXECUTION_MONITOR_SAME_TOOL_LIMIT |
5 |
Threshold for consecutive identical tool calls before mentor review |
| ExecutionMonitorTotalToolLimit | EXECUTION_MONITOR_TOTAL_TOOL_LIMIT |
10 |
Threshold for total tool calls before mentor review |
| MaxGeneralAgentToolCalls | MAX_GENERAL_AGENT_TOOL_CALLS |
100 |
Maximum tool calls for general agents (Assistant, Primary, Pentester, Coder, Installer) |
| MaxLimitedAgentToolCalls | MAX_LIMITED_AGENT_TOOL_CALLS |
20 |
Maximum tool calls for limited agents (Searcher, Enricher, etc.) |
| AgentPlanningStepEnabled | AGENT_PLANNING_STEP_ENABLED |
false |
Enable automatic task planning for specialist agents |
Usage Details
The agent supervision settings are used in pkg/providers/providers.go and pkg/providers/performer.go to configure supervision mechanisms:
-
ExecutionMonitorEnabled: Controls whether execution monitoring (mentor) is active:
buildMonitor: func() *executionMonitor { return &executionMonitor{ enabled: pc.cfg.ExecutionMonitorEnabled, sameThreshold: pc.cfg.ExecutionMonitorSameToolLimit, totalThreshold: pc.cfg.ExecutionMonitorTotalToolLimit, } } -
ExecutionMonitorSameToolLimit: Sets the threshold for identical consecutive tool calls:
// In executionMonitor.shouldInvokeMentor if emd.sameToolCount >= emd.sameThreshold { // Invoke mentor (adviser agent) for execution review }When an agent calls the same tool this many times consecutively, the execution monitor automatically invokes the mentor (adviser agent) to analyze progress and provide guidance.
-
ExecutionMonitorTotalToolLimit: Sets the threshold for total tool calls:
// In executionMonitor.shouldInvokeMentor if emd.totalCallCount >= emd.totalThreshold { // Invoke mentor (adviser agent) for execution review }When an agent makes this many total tool calls since the last mentor review, the execution monitor automatically invokes the mentor to prevent inefficient loops and provide strategic guidance.
-
MaxGeneralAgentToolCalls: Maximum iterations for general-purpose agents with full capabilities:
// In performAgentChain switch optAgentType { case pconfig.OptionsTypeAssistant, pconfig.OptionsTypePrimaryAgent, pconfig.OptionsTypePentester, pconfig.OptionsTypeCoder, pconfig.OptionsTypeInstaller: if fp.maxGACallsLimit <= 0 { maxCallsLimit = maxGeneralAgentChainIterations // fallback: 100 } else { maxCallsLimit = max(fp.maxGACallsLimit, maxAgentShutdownIterations*2) } }General agents (Assistant, Primary Agent, Pentester, Coder, Installer) are designed for complex, multi-step workflows and have a higher tool call limit to complete sophisticated tasks.
-
MaxLimitedAgentToolCalls: Maximum iterations for specialized, limited-scope agents:
// In performAgentChain default: if fp.maxLACallsLimit <= 0 { maxCallsLimit = maxLimitedAgentChainIterations // fallback: 20 } else { maxCallsLimit = max(fp.maxLACallsLimit, maxAgentShutdownIterations*2) } }Limited agents (Searcher, Enricher, Memorist, Generator, Reporter, Adviser, Reflector, Planner) are designed for focused, specific tasks and have a lower tool call limit to ensure efficient execution.
-
AgentPlanningStepEnabled: Controls automatic task planning for specialist agents:
// In flowProvider initialization planning: pc.cfg.AgentPlanningStepEnabled // Used when invoking specialist agents if fp.planning { // Generate execution plan via planner before specialist execution }When enabled, the planner (adviser in planning mode) generates a structured 3-7 step execution plan before specialist agents (Pentester, Coder, Installer) begin their work, improving task completion rates and preventing scope creep.
These settings enable:
- Automatic supervision: Mentor reviews execution patterns to detect loops and inefficiencies
- Graceful termination: Reflector guides agents to proper completion when approaching limits
- Differentiated capabilities: General agents have more autonomy for complex workflows
- Efficient execution: Limited agents stay focused on their specific scope
- Strategic planning: Automatic task decomposition for better execution quality
Supervision System Integration
The supervision settings work together as a comprehensive system:
-
Execution Monitoring (via ExecutionMonitor settings):
- Detects repetitive patterns (same tool called N times)
- Detects excessive exploration (total tools called N times)
- Automatically invokes mentor for guidance and correction
- Resets counters after mentor review
-
Tool Call Limits (via MaxGeneralAgentToolCalls and MaxLimitedAgentToolCalls):
- Prevents runaway executions with hard limits
- Invokes reflector for graceful termination near limit
- Different limits for different agent capabilities
- Ensures system stability and resource efficiency
-
Task Planning (via AgentPlanningStepEnabled):
- Generates structured execution plans before specialist work
- Prevents scope creep and maintains focus
- Improves success rates for complex tasks
- Provides clear verification points
Recommended Settings
-
Production Environment:
ExecutionMonitorEnabled: false ExecutionMonitorSameToolLimit: 5 ExecutionMonitorTotalToolLimit: 10 MaxGeneralAgentToolCalls: 100 MaxLimitedAgentToolCalls: 20 AgentPlanningStepEnabled: falseDefault settings provide stable execution without beta features.
-
High-Complexity Workflows:
ExecutionMonitorEnabled: false ExecutionMonitorSameToolLimit: 7 ExecutionMonitorTotalToolLimit: 15 MaxGeneralAgentToolCalls: 150 MaxLimitedAgentToolCalls: 30 AgentPlanningStepEnabled: falseIncreased limits for tasks requiring extensive exploration and iteration.
-
Resource-Constrained Environment:
ExecutionMonitorEnabled: false ExecutionMonitorSameToolLimit: 3 ExecutionMonitorTotalToolLimit: 7 MaxGeneralAgentToolCalls: 50 MaxLimitedAgentToolCalls: 15 AgentPlanningStepEnabled: falseTighter limits to reduce resource usage.
-
Debugging Mode:
ExecutionMonitorEnabled: false MaxGeneralAgentToolCalls: 200 MaxLimitedAgentToolCalls: 50 AgentPlanningStepEnabled: falseDisabled supervision for debugging to observe natural agent behavior.
Observability Settings
These settings control the observability and monitoring capabilities, including telemetry and trace collection for system performance and debugging.
Telemetry
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| TelemetryEndpoint | OTEL_HOST |
(none) | Endpoint for OpenTelemetry data collection |
Langfuse
| Option | Environment Variable | Default Value | Description |
|---|---|---|---|
| LangfuseBaseURL | LANGFUSE_BASE_URL |
(none) | Base URL for Langfuse API |
| LangfuseProjectID | LANGFUSE_PROJECT_ID |
(none) | Project ID for Langfuse |
| LangfusePublicKey | LANGFUSE_PUBLIC_KEY |
(none) | Public key for Langfuse API |
| LangfuseSecretKey | LANGFUSE_SECRET_KEY |
(none) | Secret key for Langfuse API |
Usage Details
The observability settings are used in main.go and the observability package to initialize monitoring systems:
-
Telemetry Configuration: Sets up OpenTelemetry for metrics, logs, and traces:
// Check if telemetry is configured if cfg.TelemetryEndpoint == "" { return nil, ErrNotConfigured } // Create telemetry client with endpoint otelclient, err := obs.NewTelemetryClient(ctx, cfg) -
Langfuse Configuration: Configures Langfuse for LLM operation monitoring:
// Check if Langfuse is configured if cfg.LangfuseBaseURL == "" { return nil, ErrNotConfigured } // Configure Langfuse client langfuse.WithBaseURL(cfg.LangfuseBaseURL), langfuse.WithPublicKey(cfg.LangfusePublicKey), langfuse.WithSecretKey(cfg.LangfuseSecretKey), langfuse.WithProjectID(cfg.LangfuseProjectID), -
Integration in Application: Used in
main.goto initialize observability:lfclient, err := obs.NewLangfuseClient(ctx, cfg) if err != nil && !errors.Is(err, obs.ErrNotConfigured) { log.Fatalf("Unable to create langfuse client: %v\n", err) } otelclient, err := obs.NewTelemetryClient(ctx, cfg) if err != nil && !errors.Is(err, obs.ErrNotConfigured) { log.Fatalf("Unable to create telemetry client: %v\n", err) } obs.InitObserver(ctx, lfclient, otelclient, []logrus.Level{ logrus.DebugLevel, logrus.InfoLevel, logrus.WarnLevel, logrus.ErrorLevel, })
These settings enable:
- Comprehensive monitoring of system performance
- LLM-specific metrics collection via Langfuse
- Tracing of requests through the system
- Centralized logging for troubleshooting
- Performance optimization based on collected metrics