This is the frontend code for ALIGN Eval, a prototype game to help you build and optimize LLM evaluators. Read about the workflow and how it was built here.
ALIGN Eval helps you build better LLM evaluators through:
- 🎮 A prototype game that makes building LLM-evaluators easy and fun
- 📊 Tools to evaluate your prompts against labeled data
- ✨ Semi-automatic optimization to improve your LLM-evaluators
- 🔄 An iterative workflow to align annotators with AI output, and AI with annotator input
While framed as a game to build LLM evaluators, you can use ALIGN Eval to craft and optimize any prompt that does binary classification!
First, run the development server:
npm run dev
# or
yarn dev
# or
pnpm dev
# or
bun dev
Open http://localhost:3000 with your browser to see the result.
- Framework: Next.js with TypeScript
- Styling: Tailwind CSS
- Fonts: Uses
next/font
to optimize and load Geist
ALIGN Eval is currently in beta. Please share constructive feedback and report bugs on GitHub or X.