Multi-agent group chat with real Letta agents. Use a modern Web GUI or a friendly CLI to run authentic multi-agent discussions with live minutes and exports.
This project implements a multi-agent group chat system using the Subjective Priority-Driven Swarm (SPDS) framework, where AI agents autonomously decide when to contribute to conversations based on internal priority calculations. Built on the Letta platform, it features real computational beings from a Letta ADE server. Available as both a CLI application and a modern web interface, it features intuitive agent selection and multiple conversation modes for rich, dynamic discussions.
π Major Update: Successfully completed conversation logic refactor (September 2025) - now featuring dynamic, context-aware agent interactions with natural conversation flow and incremental message delivery.
Key Innovation: Agents use their own LLM models to perform real subjective assessment of conversations, creating authentic computational personalities that naturally respond, agree, disagree, and build on each other's ideas. The system now uses recent conversation context instead of static topics for more relevant and engaging interactions.
π Hybrid Mode (Recommended): Two-phase conversations
-
Phase 1: All motivated agents share independent thoughts on an initial topic or question.
-
Phase 2: Agents respond to each other's ideas with rebuttals, agreements, or new insights
π₯ All-Speak Mode: Fast-paced group discussions
- All motivated agents respond in priority order
- Each agent sees previous responses within the same turn
π Sequential Mode: Traditional turn-taking
- One agent speaks per turn with fairness rotation
- Prevents any single agent from dominating
π― Pure Priority Mode: Meritocratic discussions
- The most motivated agent always speaks
- Natural leader-follower dynamics
- π€ Interactive Agent Selection: Checkbox-based UI to select computational beings from your Letta server
- π Multiple Conversation Modes: Four distinct modes for different discussion dynamics
- π Secretary Agent: AI-powered meeting documentation using real Letta agent intelligence
- π Meeting Minutes: Both formal board minutes and casual discussion notes
- π§ Real Agent Assessment: Agents use their own LLM models for authentic conversation evaluation
- π Natural Group Dynamics: Agents respond, agree, disagree, and build on each other's ideas
- β‘ Priority-Based Responses: Dynamic turn-taking based on agent motivation calculations
- πΎ Multi-Format Export: Export conversations, minutes, transcripts, and summaries
- π Dynamic Context Awareness: Agents evaluate conversation relevance using recent messages instead of static topics
- π Incremental Message Delivery: Efficient message processing with ConversationMessage architecture
- π₯οΈ Command Line Interface: Interactive terminal application with checkbox selection
- π Web Interface: Modern Bootstrap 5 web GUI with real-time WebSocket communication
- π± Responsive Design: Web interface works seamlessly on desktop and mobile devices
- π Model Diversity: Supports agents with different LLM providers (OpenAI, Anthropic, Meta, etc.)
- π Secure Authentication: Proper self-hosted Letta server integration with password authentication
- π Real-Time Assessment: Agents evaluate conversation relevance across 7 dimensions
- π¬ Human-in-the-Loop: Seamless interaction between human beings and computational beings
- β¨οΈ Live Commands: Real-time meeting management with slash commands
- π WebSocket Communication: Real-time updates and live agent responses in web interface
python3 -m venv .venv-web && . .venv-web/bin/activate # Most Linux distros use `python3`, other OSs may use `python`
pip install -r swarms-web/requirements.txt
cd swarms-web && python run.py # http://localhost:5002python3 -m venv .venv && . .venv/bin/activate
pip install -r requirements.txt
python3 -m spds.mainSet environment variables in a .env file:
LETTA_PASSWORD=your-server-password
LETTA_BASE_URL=http://localhost:8283
LETTA_ENVIRONMENT=SELF_HOSTED # or LETTA_CLOUDSee .env.example for complete configuration options.
If you encounter installation errors with letta-flask dependencies, try:
- Use the local shim: run web server from repo root
- Install without dependencies:
pip install --no-deps git+https://github.com/letta-ai/letta-flask.git
Pull requests welcome! Please run tests/linters before submitting.
MIT β see LICENSE.