Skip to content

Troubleshooting

Issues specific to Llamactl deployment and operation.

Configuration Issues

Invalid Configuration

Problem: Invalid configuration preventing startup

Solutions: 1. Use minimal configuration:

server:
  host: "0.0.0.0"
  port: 8080
instances:
  port_range: [8000, 9000]

  1. Check data directory permissions:
    # Ensure data directory is writable (default: ~/.local/share/llamactl)
    mkdir -p ~/.local/share/llamactl/{instances,logs}
    

Instance Management Issues

Model Loading Failures

Problem: Instance fails to start with model loading errors

Common Solutions:
- llama-server not found: Ensure llama-server binary is in PATH
- Wrong model format: Ensure model is in GGUF format
- Insufficient memory: Use smaller model or reduce context size
- Path issues: Use absolute paths to model files

Memory Issues

Problem: Out of memory errors or system becomes unresponsive

Solutions: 1. Reduce context size:

{
  "n_ctx": 1024
}

  1. Use quantized models:
  2. Try Q4_K_M instead of higher precision models
  3. Use smaller model variants (7B instead of 13B)

GPU Configuration

Problem: GPU not being used effectively

Solutions: 1. Configure GPU layers:

{
  "n_gpu_layers": 35
}

Advanced Instance Issues

Problem: Complex model loading, performance, or compatibility issues

Since llamactl uses llama-server under the hood, many instance-related issues are actually llama.cpp issues. For advanced troubleshooting:

Resources:
- llama.cpp Documentation: https://github.com/ggml/llama.cpp
- llama.cpp Issues: https://github.com/ggml/llama.cpp/issues
- llama.cpp Discussions: https://github.com/ggml/llama.cpp/discussions

Testing directly with llama-server:

# Test your model and parameters directly with llama-server
llama-server --model /path/to/model.gguf --port 8081 --n-gpu-layers 35

This helps determine if the issue is with llamactl or with the underlying llama.cpp/llama-server.

API and Network Issues

CORS Errors

Problem: Web UI shows CORS errors in browser console

Solutions: 1. Configure allowed origins:

server:
  allowed_origins:
    - "http://localhost:3000"
    - "https://yourdomain.com"

Authentication Issues

Problem: API requests failing with authentication errors

Solutions: 1. Disable authentication temporarily:

auth:
  require_management_auth: false
  require_inference_auth: false

  1. Configure API keys:

    auth:
      management_keys:
        - "your-management-key"
      inference_keys:
        - "your-inference-key"
    

  2. Use correct Authorization header:

    curl -H "Authorization: Bearer your-api-key" \
      http://localhost:8080/api/v1/instances
    

Debugging and Logs

Viewing Instance Logs

# Get instance logs via API
curl http://localhost:8080/api/v1/instances/{name}/logs

# Or check log files directly
tail -f ~/.local/share/llamactl/logs/{instance-name}.log

Enable Debug Logging

export LLAMACTL_LOG_LEVEL=debug
llamactl

Getting Help

When reporting issues, include:

  1. System information:

    llamactl --version
    

  2. Configuration file (remove sensitive keys)

  3. Relevant log output

  4. Steps to reproduce the issue