e2e
The platform for LLM evaluations and AI agent testing
Installable AI agent skills for prompt engineering developer workflows in Claude Code, Cursor, and Windsurf.
This collection is built for teams that treat prompt engineering as an engineering discipline, not ad-hoc tweaking. It covers design patterns, evaluation loops, regression checks, and versioned workflows that keep agent behavior stable as products evolve. MCP may show up as a supporting integration in some setups, but the core is repeatable prompt quality and reliable execution in production.
The platform for LLM evaluations and AI agent testing
Agentic coding framework powered by AGENTS.md — systematic, test-first workflows with quality gates for Cursor, Codex, Gemini CLI, and AI coding agents.
Agentic coding framework powered by AGENTS.md — systematic, test-first workflows with quality gates for Cursor, Codex, Gemini CLI, and AI coding agents.
Systematic workflows for AI-assisted development - Task-oriented framework with quality gates
Debug, evaluate, and monitor your LLM applications, RAG systems, and agentic workflows with comprehensive tracing, automated evaluations, and production-ready dashboards.
Debug, evaluate, and monitor your LLM applications, RAG systems, and agentic workflows with comprehensive tracing, automated evaluations, and production-ready dashboards.
f.k.a. Awesome ChatGPT Prompts. Share, discover, and collect prompts from the community. Free and open source — self-host for your organization with complete privacy.
a.k.a. Awesome ChatGPT Prompts. Share, discover, and collect prompts from the community. Free and open source — self-host for your organization with complete privacy.