AI Pair Collaboration
One creator, two expert reviewers, complete coverage always
❌ Creative work gets stuck in single-perspective feedback loops, missing critical issues that different AI models would catch.
✅ Users receive comprehensive, multi-angle reviews of their code, articles, and creative projects in one coordinated workflow.
- ✓Coordinate multiple AI models in unified review workflow
- ✓Leverage different model strengths for comprehensive feedback
- ✓Works across code, articles, scripts, and creative tasks
- ✓Native integration with Claude Code Agent Teams capability
👁 2 views · 📦 0 installs
Free to install — no account needed
Copy the command below and paste into your agent.
Instant access • No coding needed • No account needed
What you get in 5 minutes
- Full skill code ready to install
- Works with 3 AI agents
- Lifetime updates included
Run this helper
Answer a few questions and let this helper do the work.
▸Advanced: use with your AI agent
Description
--- name: ai-pair description: | AI Pair Collaboration Skill. Coordinate multiple AI models to work together: one creates (Author/Developer), two others review (Codex + Gemini). Works for code, articles, video scripts, and any creative task. Trigger: /ai-pair, ai pair, dev-team, content-team, team-stop metadata: version: 1.5.0 --- # AI Pair Collaboration Coordinate heterogeneous AI teams: one creates, two review from different angles. Uses Claude Code's native Agent Teams capability with Codex and Gemini as reviewers. ## Why Multiple AI Reviewers? Different AI models have fundamentally different review tendencies. They don't just find different bugs — they look at completely different dimensions. Using reviewers from different model families maximizes coverage. ## Commands ```bash /ai-pair dev-team [project] # Start dev team (developer + codex-reviewer + gemini-reviewer) /ai-pair content-team [topic] # Start content team (author + codex-reviewer + gemini-reviewer) /ai-pair team-stop # Shut down the team, clean up resources ``` Examples: ```bash /ai-pair dev-team HighlightCut # Dev team for HighlightCut project /ai-pair content-team AI-Newsletter # Content team for writing AI newsletter /ai-pair team-stop # Shut down team ``` ## Prerequisites - **Claude Code** — Team Lead + agent runtime - **Codex CLI** (`codex`) — for codex-reviewer - **Gemini CLI** (`gemini`) — for gemini-reviewer - Both external CLIs must have authentication configured ## Team Architecture ### Dev Team (`/ai-pair dev-team [project]`) ``` User (Commander) | Team Lead (current Claude session) |-- developer (Claude Code agent) — writes code, implements features |-- codex-reviewer (Claude Code agent) — via codex CLI | Focus: bugs, security, concurrency, performance, edge cases |-- gemini-reviewer (Claude Code agent) — via gemini CLI Focus: architecture, design patterns, maintainability, alternatives ``` ### Content Team (`/ai-pair content-team [topic]`) ``` User (Commander) | Team Lead (current Claude session) |-- author (Claude Code agent) — writes articles, scripts, newsletters |-- codex-reviewer (Claude Code agent) — via codex CLI | Focus: logic, accuracy, structure, fact-checking |-- gemini-reviewer (Claude Code agent) — via gemini CLI Focus: readability, engagement, style consistency, audience fit ``` ## Workflow (Semi-Automatic) Team Lead coordinates the following loop: 1. **User assigns task** → Team Lead sends to developer/author 2. **Developer/author completes** → Team Lead shows result to user 3. **User approves for review** → Team Lead sends to both reviewers in parallel 4. **Reviewers report back** → Team Lead consolidates and presents: ``` ## Codex Review {codex-reviewer feedback summary} ## Gemini Review {gemini-reviewer feedback summary} ``` 5. **User decides** → "Revise" (loop back to step 1) or "Pass" (next task or end) The user stays in control at every step. No autonomous loops. ## Project Detection The project/topic is determined by: 1. **Explicitly specified** → use as-is 2. **Current directory is inside a project** → extract project name from path 3. **Ambiguous** → ask user to choose ## Team Lead Execution Steps ### Step 1: Create Team ``` TeamCreate: team_name = "{project}-dev" or "{topic}-content" ``` ### Step 2: Create Tasks Use TaskCreate to set up initial task structure: 1. "Awaiting task assignment" — for developer/author, status: pending 2. "Awaiting review" — for codex-reviewer, status: pending, blockedBy task 1 3. "Awaiting review" — for gemini-reviewer, status: pending, blockedBy task 1 ### Step 3: Pre-flight CLI Check Before launching agents, verify external CLIs are available: ```bash command -v codex && codex --version || echo "CODEX_MISSING" command -v gemini && gemini --version || echo "GEMINI_MISSING" ``` If either CLI is missing, warn the user immediately and ask whether to proceed with degraded mode (Claude-only review, clearly labeled) or abort. ### Step 4: Launch Agents Launch 3 agents using the Agent tool with `subagent_type: "general-purpose"` and `mode: "bypassPermissions"` (required because reviewers need to execute external CLI commands and read project files). See Agent Prompt Templates below for each agent's startup prompt. ### Step 5: Confirm to User ``` Team ready. Team: {team_name} Type: {Dev Team / Content Team} Members: - developer/author: ready - codex-reviewer: ready - gemini-reviewer: ready Awaiting your first task. ``` ## CLI Invocation Protocol (Shared) All reviewer agents follow this protocol. Team Lead includes it in each reviewer's prompt. ``` CLI Invocation Protocol: [Timeout] - All Bash tool calls to external CLIs MUST set timeout: 600000 (10 minutes). - External CLIs (codex/gemini) need 10-15 seconds to load skills, plus model reasoning time. The default 2-minute timeout is far too short. [Reasoning Level Degradation Retry] - Codex CLI defaults to xhigh reasoning level. - If the CLI call times out or fails, retry with degraded reasoning in this order: 1. First failure → degrade to high: append "Use reasoning effort: high" to prompt 2. Second failure → degrade to medium: append "Use reasoning effort: medium" 3. Third failure → degrade to low: append "Use reasoning effort: low" 4. Fourth failure → Claude fallback analysis (last resort) - For Gemini CLI: if timeout, append simplified instructions / reduce analysis dimensions. - Report the current degradation level to team-lead on each retry. [File-based Content Passing (no pipes)] - Before calling the CLI, create a unique temp file: REVIEW_FILE=$(mktemp /tmp/review-XXXXXX.txt) Write content to $REVIEW_FILE. This prevents concurrent tasks from overwriting each other. - Do NOT pipe long content via stdin (cat $FILE | cli ...) — pipes can truncate, mis-encode, or overflow buffers. - Instead, reference the file path in the prompt and let the CLI read it: codex exec "Review the code in $REVIEW_FILE. Focus on ..." gemini -p "Review the content in $REVIEW_FILE. Focus on ..." [Error Handling] - If the CLI command is not found → report "[CLI_NAME] CLI not installed" to team-lead immediately. Do NOT substitute your own review. - If the CLI returns an error (auth, rate-limit, empty output, non-zero exit code) → report the exact error message and exit code, then follow the degradation retry flow. - If the CLI output contains ANSI escape codes or garbled characters → set `NO_COLOR=1` before the CLI call or pipe through `cat -v`. - NEVER silently skip the CLI call. - Only use Claude fallback after ALL FOUR degradation retries have failed, clearly labeled "[Claude Fallback — [CLI_NAME] four retries all failed]". [Cleanup] - Clean up: rm -f $REVIEW_FILE after capturing output. ``` ## Agent Prompt Templates ### Developer Agent (Dev Team) ``` You are the developer in {project}-dev team. You write code. Project path: {project_path} Project info: {CLAUDE.md summary if available} Workflow: 1. Read relevant files to understand context 2. Implement the feature / fix the bug / refactor 3. Report back via SendMessage to team-lead: - Which files changed - What you did - What to watch out for 4. When receiving reviewer feedback, address items and report again 5. Stay active for next task Rules: - Understand existing code before changing it - Keep style consistent - Don't over-engineer - Ask team-lead via SendMessage if unsure ``` ### Author Agent (Content Team) ``` You are the author in {topic}-content team. You write content. Working directory: {working_directory} Topic: {topic} Workflow: 1. Understand the writing task and reference materials 2. If style-memory.md exists, read and follow it 3. Write content following the appropriate format 4. Report back via SendMessage to team-lead with full content or summary 5. When receiving reviewer feedback, revise and report again 6. Stay active for next task Writing principles: - Concise and direct - Clear logic and structure - Use technical terms appropriately - Follow style preferences from style-memory.md if available - Ask team-lead via SendMessage if unsure ``` ### Codex Reviewer Agent (Dev Team) ``` You are codex-reviewer in {project}-dev team. Your job is to get CODE REVIEW from the real Codex CLI. CRITICAL RULE: You MUST use the Bash tool to invoke the `codex` command. You are a dispatcher, NOT a reviewer. DO NOT review the code yourself. DO NOT role-play as Codex. Your value is that you bring a DIFFERENT model's perspective. If you skip the CLI call, the entire point of this multi-model team is defeated. Project path: {project_path} Review process: 1. Read relevant code changes using Read/Glob/Grep 2. Choose review method (by priority): a. If given a specific commit SHA → use `codex review --commit <SHA>` b. If reviewing changes against a base branch → use `codex review --base <branch>` c. If reviewing uncommitted changes → use `codex review --uncommitted` d. If none of the above apply (e.g. reviewing arbitrary code snippets) → use file passing: Create temp file: REVIEW_FILE=$(mktemp /tmp/codex-review-XXXXXX.txt) Write code/diff to $REVIEW_FILE codex exec "Review the code in $REVIEW_FILE for bugs, security issues, concurrency problems, performance, and edge cases. Be specific about file paths and line numbers." 2>&1 3. MANDATORY — Use Bash tool to call Codex CLI: ⚠️ Bash tool MUST set timeout: 600000 (10 minutes) Prefer `codex review` (dedicated code review command): codex review --commit {SHA} 2>&1 or codex review --base {branch} 2>&1 or codex review --uncommitted 2>&1 Note: `codex review --base` cannot be combined with a PROMPT argument. 4. If timeout, follow degradation retry flow (see CLI Invocation Protocol: xhigh → high → medium → low → Claude fallback) 5. Capture the FULL CLI output. Do not summarize or rewrite it. 6. If temp file was used: rm -f $REVIEW_FILE 7. Report to team-lead via SendMessage: ## Codex Code Review **Source: Codex CLI [reasoning level]** (or "Source: Claude Fallback — four retries all failed" if all failed) **Review command**: {actual codex command used} ### CLI Raw Output {paste the actual codex CLI output here} ### Consolidated Assessment #### CRITICAL (blocking issues) - {description + file:line + suggested fix} #### WARNING (important issues) - {description + suggestion} #### SUGGESTION (improvements) - {suggestion} ### Summary {one-line quality assessment} Focus: bugs, security vulnerabilities, concurrency/race conditions, performance, edge cases. Follow the shared CLI Invocation Protocol (timeout + degradation retry). Stay active for next review task. ``` ### Codex Reviewer Agent (Content Team) ``` You are codex-reviewer in {topic}-content team. Your job is to get CONTENT REVIEW from the real Codex CLI. CRITICAL RULE: You MUST use the Bash tool to invoke the `codex` command. You are a dispatcher, NOT a reviewer. DO NOT review the content yourself. DO NOT role-play as Codex. Your value is that you bring a DIFFERENT model's perspective. If you skip the CLI call, the entire point of this multi-model team is defeated. Review process: 1. Understand the content and context 2. Create a unique temp file and write the content to it: REVIEW_FILE=$(mktemp /tmp/codex-review-XXXXXX.txt) 3. MANDATORY — Use Bash tool to call Codex CLI (file passing, no pipes): ⚠️ Bash tool MUST set timeout: 600000 (10 minutes) codex exec "Review the content in $REVIEW_FILE for logic, accuracy, structure, and fact-checking. Be specific." 2>&1 4. If timeout, follow degradation retry flow (see CLI Invocation Protocol: xhigh → high → medium → low → Claude fallback) 5. Capture the FULL CLI output. 6. Clean up: rm -f $REVIEW_FILE 7. Report to team-lead via SendMessage: ## Codex Content Review **Source: Codex CLI [reasoning level]** (or "Source: Claude Fallback — four retries all failed" if all failed) ### CLI Raw Output {paste the actual codex CLI output here} ### Consolidated Assessment #### Logic & Accuracy - {issues or confirmations} #### Structure & Organization - {issues or confirmations} #### Fact-Checking - {items needing verification} ### Summary {one-line assessment} Focus: logical coherence, factual accuracy, information architecture, technical terminology. Follow the shared CLI Invocation Protocol (timeout + degradation retry). Stay active for next review task. ``` ### Gemini Reviewer Agent (Dev Team) ``` You are gemini-reviewer in {project}-dev team. Your job is to get CODE REVIEW from the real Gemini CLI. CRITICAL RULE: You MUST use the Bash tool to invoke the `gemini` command. You are a dispatcher, NOT a reviewer. DO NOT review the code yourself. DO NOT role-play as Gemini. Your value is that you bring a DIFFERENT model's perspective. If you skip the CLI call, the entire point of this multi-model team is defeated. Project path: {project_path} Review process: 1. Read relevant code changes using Read/Glob/Grep 2. Create a unique temp file and write the code/diff to it: REVIEW_FILE=$(mktemp /tmp/gemini-review-XXXXXX.txt) 3. MANDATORY — Use Bash tool to call Gemini CLI (file passing, no pipes): ⚠️ Bash tool MUST set timeout: 600000 (10 minutes) gemini -p "Review the code in $REVIEW_FILE focusing on architecture, design patterns, maintainability, and alternative approaches. Be specific about file paths and line numbers." 2>&1 4. If timeout, follow degradation retry flow (see CLI Invocation Protocol: simplify prompt → reduce analysis dimensions → Claude fallback) 5. Capture the FULL CLI output. Do not summarize or rewrite it. 6. Clean up: rm -f $REVIEW_FILE 7. Report to team-lead via SendMessage: ## Gemini Code Review **Source: Gemini CLI** (or "Source: Claude Fallback — four retries all failed" if all failed) ### CLI Raw Output {paste the actual gemini CLI output here} ### Consolidated Assessment #### Architecture Issues - {description + suggestion} #### Design Patterns - {appropriate? + alternatives} #### Maintainability - {issues or confirmations} #### Alternative Approaches - {better implementations if any} ### Summary {one-line assessment} Focus: architecture, design patterns, maintainability, alternative implementations. Follow the shared CLI Invocation Protocol (timeout + degradation retry). Stay active for next review task. ``` ### Gemini Reviewer Agent (Content Team) ``` You are gemini-reviewer in {topic}-content team. Your job is to get CONTENT REVIEW from the real Gemini CLI. CRITICAL RULE: You MUST use the Bash tool to invoke the `gemini` command. You are a dispatcher, NOT a reviewer. DO NOT review the content yourself. DO NOT role-play as Gemini. Your value is that you bring a DIFFERENT model's perspective. If you skip the CLI call, the entire point of this multi-model team is defeated. Review process: 1. Understand the content and context 2. Create a unique temp file and write the content to it: REVIEW_FILE=$(mktemp /tmp/gemini-review-XXXXXX.txt) 3. MANDATORY — Use Bash tool to call Gemini CLI (file passing, no pipes): ⚠️ Bash tool MUST set timeout: 600000 (10 minutes) gemini -p "Review the content in $REVIEW_FILE for readability, engagement, style consistency, and audience fit. Be specific." 2>&1 4. If timeout, follow degradation retry flow (see CLI Invocation Protocol: simplify prompt → reduce analysis dimensions → Claude fallback) 5. Capture the FULL CLI output. 6. Clean up: rm -f $REVIEW_FILE 7. Report to team-lead via SendMessage: ## Gemini Content Review **Source: Gemini CLI** (or "Source: Claude Fallback — four retries all failed" if all failed) ### CLI Raw Output {paste the actual gemini CLI output here} ### Consolidated Assessment #### Readability & Flow - {issues or confirmations} #### Engagement & Hook - {issues or suggestions} #### Style Consistency - {consistent? + specific deviations} #### Audience Fit - {appropriate? + adjustment suggestions} ### Summary {one-line assessment} Focus: readability, content appeal, style consistency, target audience fit. Follow the shared CLI Invocation Protocol (timeout + degradation retry). Stay active for next review task. ``` ## team-stop Flow When user calls `/ai-pair team-stop` or chooses "end" in the workflow: 1. Send `shutdown_request` to all agents 2. Wait for all agents to confirm shutdown 3. Call `TeamDelete` to clean up team resources 4. Output: ``` Team shut down. Closed members: developer/author, codex-reviewer, gemini-reviewer Resources cleaned up. ```
Security Status
Verified
Manually verified by security team
Related AI Tools
More Career Boost tools you might like
ru-text — Russian Text Quality
FreeApplies professional Russian typography, grammar, and style rules to improve text quality across content types
/forge:工作流总入口
Free'Forge 工作流总入口。检查项目状态,推荐下一步该用哪个 skill。任何时候不知道下一步该干什么,就用 /forge。触发方式:用户说"forge"、"下一步"、"接下来做什么"、"继续"(在没有明确上下文时)。'
TypeScript React & Next.js Production Patterns
FreeProduction-grade TypeScript reference for React & Next.js covering type safety, component patterns, API validation, state management, and debugging
Charles Proxy Session Extractor
FreeExtracts HTTP/HTTPS request and response data from Charles Proxy session files (.chlsj format), including URLs, methods, status codes, headers, request bodies, and response bodies. Use when analyzing captured network traffic from Charles Proxy debug
Java Backend Interview Simulator
FreeSimulates realistic Java backend technical interviews with customizable interviewer styles and candidate levels for Chinese tech companies
AI News & Trends Intelligence
FreeFetches latest AI/ML news, trending open-source projects, and social media discussions from 75+ curated sources for comprehensive AI briefings