A lightweight kit for quickly standing up LLM evals. It includes a simple web UI, a CLI, and a pluggable API layer that works with local models (Ollama) and major cloud providers (OpenAI, AWS Bedrock) ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results
Feedback