Thank you for your interest in contributing to AI Agents Reality Check! This project aims to provide rigorous, mathematically sound benchmarking of AI agent architectures.
- Use GitHub Issues with the "bug" label
- Include system information and reproduction steps
- Attach benchmark output logs when relevant
- Use GitHub Issues with the "enhancement" label
- Explain the use case and expected behavior
- Consider mathematical rigor and architectural focus
- New agent architectures for benchmarking
- Additional statistical analysis methods
- Academic validation studies
- Performance optimization improvements
git clone https://github.com/Cre4T3Tiv3/ai-agents-reality-check
cd ai-agents-reality-check
make install-dev- Use
blackfor formatting:make format - Type hints required for all functions
- Docstrings for all public methods
- 90%+ test coverage for new features
- Statistical methods must be academically sound
- Include confidence intervals for performance metrics
- Document assumptions and limitations
- Provide references for statistical techniques
- Unit tests for all new functionality
- Integration tests for benchmark pipelines
- Schema validation for data structures
- Performance regression tests
- Fork and Branch: Create feature branch from
main - Implement: Follow code standards and testing requirements
- Test: Run
make checkand ensure all tests pass - Document: Update relevant documentation
- Submit: Create PR with clear description and context
- Mathematical accuracy and statistical validity
- Code quality and test coverage
- Documentation completeness
- Backward compatibility
- Performance impact assessment
- Be respectful and constructive
- Focus on technical merit and empirical evidence
- Welcome newcomers and provide helpful feedback
- Maintain the ethos: challenge assumptions with data
- GitHub Discussions for general questions
- GitHub Issues for specific bugs or features
- Email ByteStack Labs for research collaborations