An automated agent for evaluating repository readiness for GitHub Copilot usage. This tool analyzes repositories across multiple dimensions to determine how effectively GitHub Copilot can assist developers.
GitHub Copilot works best when repositories follow best practices for structure, documentation, and tooling. This agent provides a comprehensive assessment and actionable recommendations to optimize your repository for Copilot usage.
The agent assesses eight key areas, each weighted based on importance to Copilot effectiveness:
- Presence of build configuration files
- Clear build instructions
- Dependency management
- CI/CD integration
- Absence of environment-specific hardcoded values
- Clear application entry points
- Runtime configuration documentation
- Environment variable templates
- Container/orchestration configurations
- Launch configurations for IDEs
- Test framework configuration
- Test file organization and naming
- Test execution documentation
- Separation of test types (unit/integration)
- CI test automation
- Logical project structure
- Consistent naming conventions
- Appropriate file sizes
- Type annotations (where applicable)
- Code style/linting configuration
- Comprehensive README
- Architecture documentation
- Contributing guidelines
- API documentation
- Inline code comments
- Presence of
.github/copilot-instructions.md - Targeted instruction files in
.github/instructions/ - Content quality and comprehensiveness
- Copilot's actual understanding of instructions
- Specificity of coding standards and tech stack guidance
- Presence of custom agents in
.github/agents/ - Proper YAML frontmatter configuration
- Clear agent descriptions and purposes
- Copilot's ability to recognize and understand agents
- Presence of reusable skills in
.github/skills/ - Proper
SKILL.mdfile structure - Clear skill descriptions and use cases
- Copilot's ability to identify and utilize skills
Copilot Readiness (30 points): Direct testing of Copilot CLI's ability to understand build, run, and test processes
Bonus Points (up to +15): GitHub Actions, Dependabot, security scanning, EditorConfig, TypeScript
Maximum Score: 165 points (135 base + 30 Copilot Readiness)
- A (90-100%): Excellent - Copilot will thrive (148-165 points)
- B (80-89%): Good - Minor improvements recommended (132-147 points)
- C (70-79%): Fair - Several enhancements needed (115-131 points)
- D (60-69%): Poor - Significant work required (99-114 points)
- F (0-59%): Failing - Major restructuring needed (0-98 points)
Build and run the assessment tool:
# Navigate to the project directory
cd repoReadiness
# Build the project
dotnet build
# Assess a repository
dotnet run -- "C:\path\to\your\repository"
# With verbose output
dotnet run -- "C:\path\to\your\repository" --verbose
# Assess current directory
dotnet run -- "."For a manual, in-depth review:
- Open
repo-readiness-agent.md - Review each criterion systematically
- Check off applicable items
- Calculate scores for each category
- Document findings and recommendations
The agent generates a detailed markdown report containing:
- Overall grade and score
- Category-by-category breakdown
- Identified strengths and weaknesses
- Prioritized recommendations
- Executive summary and conclusion
# Repository Readiness Report
**Repository:** my-awesome-app
**Date:** 2026-01-26
**Overall Grade:** B (84/100)
## Detailed Scores
- Build Capability: 22/25
- Run Capability: 18/20
- Test Capability: 15/20
- Code Understanding: 17/20
- Documentation Quality: 12/15
## Priority Actions
1. Add .env.example file
2. Configure test framework
3. Add API documentation
...
✅ Content-aware analysis - Tests Copilot's actual understanding of your instructions, agents, and skills
✅ Automated scanning - Analyzes repositories automatically with deep content inspection
✅ Comprehensive coverage - Evaluates 50+ criteria across 8 dimensions
✅ Copilot CLI integration - Uses real Copilot responses to validate readiness
✅ Actionable insights - Specific recommendations based on what Copilot can and cannot understand
✅ Bonus point detection - Rewards best practices like CI/CD and TypeScript
✅ Detailed reporting - Professional markdown reports with Copilot understanding analysis
✅ Multi-language support - Recognizes patterns for JavaScript, Python, Java, C#, Go, Rust, and more
- .NET SDK 6.0 or higher (for C# version)
- PowerShell 5.1 or higher (for PowerShell version)
- GitHub Copilot CLI (recommended for full assessment including content understanding tests)
- Read access to the target repository
For the most accurate assessment, install GitHub Copilot CLI:
# Install via GitHub CLI
gh extension install github/gh-copilot
# Or install standalone CLI from:
# https://docs.github.com/en/copilot/github-copilot-in-the-cliWithout Copilot CLI, the tool will perform static analysis only and won't be able to test whether Copilot actually understands your custom instructions, agents, and skills.
Your repository is well-prepared for Copilot. Focus on maintaining standards and addressing minor gaps.
Good foundation, but several improvements needed. Follow priority actions to optimize Copilot effectiveness.
Significant work required. Use this assessment as a roadmap for modernization and best practices adoption.
Based on assessments and the latest Copilot features (2025-2026), these practices maximize Copilot effectiveness:
- Clear structure: Organize code logically with meaningful names
- Comprehensive README: Document purpose, setup, and usage
- Build automation: Provide simple, documented build commands
- Test coverage: Include tests that demonstrate expected behavior
- Type information: Use TypeScript, type hints, or interface definitions
- Consistent style: Apply linting and formatting tools
- Environment templates: Provide
.env.examplefiles - CI/CD: Automate builds and tests
- Custom instructions: Add
.github/copilot-instructions.mdwith:- Project overview and purpose
- Tech stack details
- Coding standards and conventions
- Security practices
- Testing requirements
- Custom agents: Create specialized agents in
.github/agents/for:- Code review workflows
- Testing automation
- Documentation generation
- Architecture planning
- Agent skills: Define reusable skills in
.github/skills/for:- Common code transformations
- Project-specific patterns
- Deployment workflows
The tool now evaluates whether Copilot actually understands your instructions by:
- Asking Copilot to extract coding standards from your instructions
- Testing if Copilot can identify your tech stack
- Verifying Copilot recognizes custom agents and skills
- Scoring based on response quality, not just file presence
This agent can be extended to evaluate additional criteria. Consider contributing checks for:
- Security scanning configurations
- Performance monitoring setup
- Accessibility standards
- License compliance
- Dependency freshness
- Code complexity metrics
Version: 2.0 - Content-Aware Assessment
Last Updated: 2026-01-27
- ✨ Custom Instructions Analysis: Deep content analysis of
.github/copilot-instructions.mdand targeted instruction files - ✨ Custom Agents Detection: Evaluates
.github/agents/with YAML validation and Copilot recognition testing - ✨ Agent Skills Assessment: Checks
.github/skills/structure and Copilot's ability to utilize them - ✨ Content Understanding Tests: Uses Copilot CLI to verify actual comprehension, not just file presence
- ✨ Enhanced Scoring: Expanded to 165 points with dedicated categories for modern Copilot features
- ✨ Quality-Based Scoring: Rewards well-structured, specific instructions over generic templates
This assessment agent is provided as-is for evaluating repository quality and Copilot readiness.