-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Description
Describe the feature or problem you'd like to solve
Currently, GitHub Copilot CLI operates with a single AI agent per interaction, requiring users to manually orchestrate complex development workflows across multiple specialized roles (architecture, product management, development, research, etc.). This creates friction in end-to-end project development and requires constant context switching. Proposed Solution: Introduce a multi-agent workflow system where users can configure a "team" of specialized AI agents, each with distinct roles, expertise, and toolsets. These agents would work sequentially and collaboratively on a project from ideation to implementation, with configurable approval gates for user oversight.
Proposed solution
-
Reduced Context Switching: Users submit an idea once and let the AI team handle the complete workflow autonomously;
-
Specialized Expertise: Each agent focuses on its domain (architecture, product discovery, development, testing) with role-specific tools and knowledge;
-
Quality Assurance: Built-in approval gates ensure users maintain control while minimizing manual intervention;
-
Scalable Complexity: Handles enterprise-grade projects that require multiple perspectives and skill sets;
-
Learning & Transparency: Users can observe how different roles approach problems, improving their own development practices;
-
Time Efficiency: Parallel/sequential task execution reduces time from concept to working code;
This would be a great tool for a monorepo, which developers can use as a starting point for their projects. Additionally, this approach mitigates context window exhaustion by distributing the workload—each agent maintains a focused context limited to its specific task and relevant outputs from previous agents, rather than a single agent trying to juggle the entire project history, architecture decisions, implementation details, and testing requirements simultaneously.
Example prompts or workflows
Example 1 (New Feature Development):
copilot team create feature-workflow \
--agents "client,product-researcher,architect,developer,tester" \
--approval-gates "after:client,after:architect"
copilot team run feature-workflow \
--input "Build a user authentication system with OAuth support"
Flow:
- Client Agent: Clarifies requirements, creates user stories → [User Approval]
- Product Researcher: Searches web for OAuth best practices, security standards (tool: web_search)
- Architect: Designs system architecture, database schema
- Developer: Implements code → [User Approval]
- Tester: Generates test cases, runs validation
Example 2 (Performance and security audit):
copilot team run audit-workflow \
--input "Analyze our Express.js API for security vulnerabilities and performance bottlenecks"
Flow:
- Security Auditor: Scans for OWASP vulnerabilities
- Performance Analyst: Profiles API endpoints, identifies N+1 queries
- Architect: Proposes refactoring strategy
- Developer: Implements fixes incrementally
Example 3 (Migration Project):
copilot team run docs-workflow \
--input "Generate comprehensive API documentation from our codebase"
Flow:
- Code Analyst: Extracts endpoints, types, schemas
- Technical Writer: Generates clear documentation with examples
- Developer: Adds inline code comments
- Reviewer: Validates accuracy, suggests improvements
Additional context
Suggested Agent Configuration Structure:
# .github/copilot-team.yml
teams:
feature-workflow:
agents:
- role: client
model: "gpt-4o"
system_prompt: "Act as a client clarifying requirements..."
tools: []
- role: product-researcher
model: "claude-sonnet-4"
system_prompt: "Research market trends and technologies..."
tools: [web_search, github_search]
- role: architect
model: "gpt-4o"
system_prompt: "Design scalable system architecture..."
tools: [diagram_generator, dependency_analyzer]
- role: developer
model: "gemini-3-pro"
system_prompt: "Implement features following best practices..."
tools: [code_generator, file_editor, terminal]
- role: tester
model: "claude-sonnet-4"
system_prompt: "Create comprehensive test suites..."
tools: [test_generator, coverage_analyzer]
approval_gates:
- after: client
message: "Review requirements before proceeding?"
- after: architect
message: "Approve architecture design?"
Key features:
- Tool Isolation: Each agent has access only to its designated tools
- Context Passing: Agents can reference outputs from previous agents
- Approval Gates: Configurable checkpoints for user input
- Observability: Real-time view of which agent is active and what it's doing
- Rollback: Ability to restart from any approval gate
- Model Selection per Agent: Choose the most suitable AI model for each role (e.g., GPT-4o for architecture, Claude for research, Gemini for development) to optimize for each agent's strengths.
- Templates: Pre-built team configurations for common workflows
Technical Considerations:
- Leverage existing Copilot Chat API with role-based system prompts
- Store team configurations in .github/copilot-team.yml
- Use structured output formats (JSON/YAML) for inter-agent communication
- Implement timeout mechanisms to prevent infinite loops