goLLM provides a command-line interface for common tasks. All commands support the --help flag for more information.
| Option | Description |
|---|---|
--config PATH |
Path to configuration file |
--verbose |
Enable verbose output |
--quiet |
Suppress non-essential output |
--version |
Show version and exit |
gollm init [PATH]Initialize a new goLLM project.
Options:
--force: Overwrite existing configuration--defaults: Use default configurationExamples:
# Initialize in current directory
gollm init
# Initialize in specific directory
gollm init my_project
gollm validate [PATH]Validate Python files.
Options:
--fix: Automatically fix fixable issues--rules RULES: Comma-separated list of rule IDs to run--exclude RULES: Comma-separated list of rule IDs to excludeExamples:
# Validate current directory
gollm validate
# Validate specific file
gollm validate src/main.py
# Validate with specific rules
gollm validate --rules=line-length,function-length
gollm generate PROMPTGenerate code using LLM with quality validation.
Options:
-o, --output FILE: Output file or directory path--critical: Mark as high priority task--no-todo: Skip creating a TODO item-f, --fast: Use fast mode with minimal validation-i, --iterations INT: Number of generation iterations (default: 3)--model MODEL: Override default model--temperature FLOAT: Sampling temperature (0.0-2.0)--max-tokens INT: Maximum number of tokens to generateExamples:
# Generate code with prompt
gollm generate "Create a Flask web server"
# Save to file
gollm generate "Create a Flask web server" -o app.py
# Use fast mode for quicker results
gollm generate "Create a Flask web server" --fast
# Specify number of iterations
gollm generate "Create a Flask web server" --iterations 2
gollm directDirect API commands for fast LLM access with minimal processing.
gollm direct generate PROMPTGenerate text using direct API access without validation pipeline.
Options:
-m, --model MODEL: Model to use (default: deepseek-coder:1.3b)-t, --temperature FLOAT: Sampling temperature (0.0-1.0)--max-tokens INT: Maximum tokens to generate--api-url URL: API base URL (default: http://localhost:11434)-o, --output FILE: Output file path-f, --format FORMAT: Output format (text or json)--timeout INT: Request timeout in seconds (default: 60)Examples:
# Generate code with direct API access
gollm direct generate "Write Hello World in Python"
# Use specific model
gollm direct generate "Write Hello World in Python" --model codellama:7b
# Save output to file
gollm direct generate "Write Hello World in Python" -o hello.py
gollm direct chat PROMPTChat with LLM using direct API access without validation pipeline.
Options:
gollm direct generateExamples:
# Chat with LLM using direct API access
gollm direct chat "How do I use asyncio in Python?"
# Get JSON response
gollm direct chat "How do I use asyncio in Python?" --format json
gollm todoManage TODOs in your project.
Subcommands:
list: List all TODOsadd "DESCRIPTION": Add a new TODOcomplete ID: Mark a TODO as completepriority ID LEVEL: Set TODO priority (high/medium/low)Examples:
# List all TODOs
gollm todo list
# Add a new TODO
gollm todo add "Fix login bug"
# Complete a TODO
gollm todo complete 1
gollm changelogManage project changelog.
Subcommands:
show: Show changelogadd "CHANGE": Add a changebump VERSION: Bump version and update changelogExamples:
# Show changelog
gollm changelog show
# Add a change
gollm changelog add "Fixed critical security issue"
# Bump version
gollm changelog bump 1.2.3
gollm configManage configuration.
Subcommands:
show: Show current configurationget KEY: Get a configuration valueset KEY VALUE: Set a configuration valueedit: Edit configuration in default editorExamples:
# Show all config
gollm config show
# Get specific value
gollm config get validation.max_line_length
# Set a value
gollm config set validation.max_line_length 100
All CLI options can be set via environment variables with the GOLLM_ prefix:
# Equivalent to --verbose
export GOLLM_VERBOSE=1
# Set default model
export GOLLM_MODEL=gpt-4
| Code | Description |
|---|---|
| 0 | Success |
| 1 | General error |
| 2 | Invalid arguments |
| 3 | File not found |
| 4 | Validation failed |
| 5 | Configuration error |
Enable shell completion by adding to your shellβs rc file:
Bash:
eval "$(_GOLLM_COMPLETE=bash_source gollm)" >> ~/.bashrc
Zsh:
eval "$(_GOLLM_COMPLETE=zsh_source gollm)" >> ~/.zshrc
Fish:
gollm --completion fish > ~/.config/fish/completions/gollm.fish
See Configuration Reference for details on the configuration file format.