Remove automatic triggers (schedule and push) to ensure build parameters are explicitly specified via workflow dispatch, consistent with other CUDA/PyTorch workflows that require manual configuration. |
||
---|---|---|
.forgejo/workflows | ||
docker | ||
docs | ||
src/neuromancer | ||
templates | ||
tests | ||
.gitignore | ||
.markdownlint.yaml | ||
.pre-commit-config.yaml | ||
config.yml | ||
CONTRIBUTING.md | ||
docker-compose.yml | ||
LICENSE | ||
pyproject.toml | ||
README.md | ||
uv.lock |
Neuromancer AI Orchestrator
A self-contained AI orchestrator that provides OpenAI-compatible APIs for locally hosted models. Run voice synthesis, chat completions, and tool-enabled AI services through Docker containers with automatic resource management and zero-configuration model downloads.
Point any OpenAI-compatible application at this server and it handles the rest - no code changes needed.
What it does
Neuromancer orchestrates AI services through Docker containers, providing a unified OpenAI-compatible API for different types of models. The system handles container lifecycle, GPU memory allocation, and automatic cleanup - services start on demand, share resources efficiently, and restart automatically if something crashes.
Language models are powered by vLLM and support chat completions, tool calling, and agent workflows. You can configure different models for different purposes: lightweight chat, tool-enabled assistants, or specialised agents for complex tasks.
Voice services use Kyutai Moshi to provide zero-shot voice cloning, real-time speech-to-text, and text-to-speech. Generate voice embeddings from single audio samples and synthesise speech in real-time using your own voice or reference samples.
Documentation
Setup and deployment:
- Setup guide - Installation and production deployment
- Configuration guide - Complete
config.yml
reference - Development setup - Local development and testing
- Troubleshooting - Common issues and solutions
Using the API:
- Chat completions API - OpenAI-compatible endpoints
- Audio processing - Voice synthesis and transcription
- Tool integration - OpenAPI server setup
- Real-time API - WebSocket voice interactions
Technical details:
- Architecture overview - System design and container orchestration
Ethical use
Voice cloning technology should be used responsibly:
- Use your own voice or obtain explicit permission
- Be transparent about AI-generated content
- Consider the impact on voice actors and content creators
- Do not impersonate others without consent or create misleading content
License
This project is licensed under the Apache License 2.0 - see the LICENSE file for details.