Search
Search docs, blog posts, and ecosystem packages with citations.
Enter a query to see grounded citations.
Evaluation framework for LLM and Jido agent quality measurement
Experimental support. Open exploration with no stability guarantee.
View package metadata source →Jido Eval is an experimental framework focused on measuring and improving LLM and agent performance in Jido-based systems.
Jido Eval provides an ecosystem home for evaluation datasets, scoring methods, and repeatable quality checks.
Defines project primitives for assembling benchmark runs and collecting outputs.
Uses typed structs and CSV tooling for controlled experiment datasets.
Ships project aliases oriented around formatting, static analysis, and docs-driven iteration.