Ahnlich in Production
Ahnlich is production-ready and designed for deployment at scale. This section covers everything you need to run Ahnlich in production environments.
What You'll Learnβ
- Deployment - Docker-based deployment strategies for cloud and on-premise
- Tracing - Distributed tracing setup for observability
Architecture Overviewβ
A typical production setup consists of:
βββββββββββββββ
β Clients β
ββββββββ¬βββββββ
β
ββββββββββββββββββββββββ
β β
βΌ βΌ
βββββββββββββββ βββββββββββββββ
β Ahnlich AI ββββββββ>β Ahnlich DB β
β (Port 1370)β β (Port 1369) β
βββββββββββββββ βββββββββββββββ
- Ahnlich DB handles vector storage and similarity search
- Ahnlich AI transforms inputs (text/images) into embeddings
- Both services communicate over gRPC
Key Features for Productionβ
Persistenceβ
Both services support disk persistence to survive restarts:
- Configurable intervals for snapshots
- Automatic recovery on startup
Performanceβ
- In-memory operations for low latency
- Batch processing support
- Configurable model batch sizes
Observabilityβ
- Distributed tracing with OpenTelemetry
- Integration with Jaeger and other collectors
- Request/response logging
Scalabilityβ
- Horizontal scaling via multiple instances
- Load balancing support
- Model caching to reduce startup time
Quick Startβ
Get started with Docker Compose:
curl -O https://raw.githubusercontent.com/deven96/ahnlich/main/docker-compose.yml
docker-compose up -d
This starts both services with persistence enabled.
Next Stepsβ
- Deploy to Production - Choose your deployment platform
- Enable Tracing - Set up observability
- Review the CLI reference for configuration options