A Telegram multi-bot framework with:
- separated group bots and private bots
- multi-role collaboration across
OpenClaw / Gemini / Codex / Claude - shared task queue
- shared memory summaries
- one-command installation for
OpenClaw / Gemini CLI / Codex / Claude Code - one-command env and launchd generation
- flexible bot count scaling
This project is suitable for:
- task routing and reporting in team Telegram groups
- high-permission execution in private chat
- running multiple bots at once with clearly separated responsibilities
flowchart LR
subgraph G["Group Bots"]
OG["OpenClaw-Group<br/>routing / decomposition / status / shared memory"]
GG["Gemini-Group<br/>daily reports / analysis / reporting"]
CG["Codex-Group<br/>coding / scripts / debugging"]
end
subgraph P["Private Bots"]
OP["OpenClaw-Private<br/>personal control / private delegation"]
GP["Gemini-Private<br/>high-permission autonomous execution"]
CP["Codex-Private<br/>private coding execution"]
end
subgraph S["Shared Layer"]
Q["Task Queue"]
M["Memory Summaries"]
K["Shared Skills"]
end
OG --> Q
GG --> Q
CG --> Q
OP --> Q
GP --> Q
CP --> Q
Q --> M
Q --> K
Language:
- Chinese: README.md
- English: README.en.md
- Chinese install: INSTALL.md
- English install: INSTALL.en.md
- AI CLI setup: docs/ai-runtimes.md
- Contributing: CONTRIBUTING.md
- Security: SECURITY.md
- Code of Conduct: CODE_OF_CONDUCT.md
- FAQ: docs/faq.md
- Changelog: CHANGELOG.md
- Architecture Guide: docs/architecture.md
git clone https://github.com/ukgorclawbot-stack/telegram-multi-bot-stack.git
cd telegram-multi-bot-stack
bash ./install.sh
bash ./configure_ai_runtimes.sh
bash ./configure.sh
bash ./apply_stack.shIf you want a beginner-friendly walkthrough, read:
If you only want to preview generated files without starting services:
git clone https://github.com/ukgorclawbot-stack/telegram-multi-bot-stack.git
cd telegram-multi-bot-stack
bash ./install.sh
bash ./configure_ai_runtimes.sh
bash ./configure.sh
bash ./bootstrap_bot_stack.sh generategroup_bot.py: shared entrypoint for group and private bot rolesbot.py: compatibility entrypoint for legacy Codex direct handlingbootstrap_bot_stack.py: generates env and launchd files from a TOML stack specconfigure_stack.py: interactive configuration wizardbootstrap_bot_stack.sh: wrapper for generate/apply/export-live/migration-templateapply_stack.sh: loads local tokens and applies the stack
# Install dependencies
bash ./install.sh
# Configure the 4 AI CLIs
bash ./configure_ai_runtimes.sh
# Run interactive configuration
bash ./configure.sh
# Generate files only, do not start services
bash ./bootstrap_bot_stack.sh generate
# Generate and start services
bash ./apply_stack.sh
# Run system health checks
bash ./health_check.sh# Reverse export a live stack into a sanitized TOML file
bash ./bootstrap_bot_stack.sh export-live
# Build a migration-ready template for a fresh machine
bash ./bootstrap_bot_stack.sh migration-templateMIT